[ 441.033090] env[61906]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61906) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 441.033090] env[61906]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61906) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 441.033090] env[61906]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61906) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 441.033482] env[61906]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 441.139192] env[61906]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61906) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 441.149485] env[61906]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61906) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 441.760715] env[61906]: INFO nova.virt.driver [None req-7138ba5b-b772-495d-bc9e-bafaf7111a2a None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 441.838698] env[61906]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 441.838927] env[61906]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 441.838997] env[61906]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61906) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 444.859219] env[61906]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-320f3cdc-7df2-4d5a-a070-6405f0c85202 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 444.874927] env[61906]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61906) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 444.875108] env[61906]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-7bcf82cc-d36e-4386-90a2-67bc16f4048a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 444.903250] env[61906]: INFO oslo_vmware.api [-] Successfully established new session; session ID is ff7a3. [ 444.903250] env[61906]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.062s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 444.903250] env[61906]: INFO nova.virt.vmwareapi.driver [None req-7138ba5b-b772-495d-bc9e-bafaf7111a2a None None] VMware vCenter version: 7.0.3 [ 444.905268] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d6de793-3588-444b-87cb-787358595314 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 444.922499] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c50764-cdaf-4a51-b1c1-ee6561fbc848 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 444.928214] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99373210-81b4-4824-91a2-4ec5ebf719cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 444.934634] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97bf8528-a97c-4c13-a5d8-1a1475a653cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 444.947818] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a708a8-a444-4124-8fb1-b6f9f91289b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 444.953606] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c9e5a6-a639-453c-b6e3-6e044e8e6a8b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 444.982996] env[61906]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-8d0997a4-c65f-4f2f-b35d-43071a43f82c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 444.987743] env[61906]: DEBUG nova.virt.vmwareapi.driver [None req-7138ba5b-b772-495d-bc9e-bafaf7111a2a None None] Extension org.openstack.compute already exists. {{(pid=61906) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 444.990476] env[61906]: INFO nova.compute.provider_config [None req-7138ba5b-b772-495d-bc9e-bafaf7111a2a None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 445.494392] env[61906]: DEBUG nova.context [None req-7138ba5b-b772-495d-bc9e-bafaf7111a2a None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),c1fc5465-5ac4-4b84-8fa6-f9f4dedbb6c1(cell1) {{(pid=61906) load_cells /opt/stack/nova/nova/context.py:464}} [ 445.497044] env[61906]: DEBUG oslo_concurrency.lockutils [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 445.497200] env[61906]: DEBUG oslo_concurrency.lockutils [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 445.497877] env[61906]: DEBUG oslo_concurrency.lockutils [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 445.498326] env[61906]: DEBUG oslo_concurrency.lockutils [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Acquiring lock "c1fc5465-5ac4-4b84-8fa6-f9f4dedbb6c1" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 445.498513] env[61906]: DEBUG oslo_concurrency.lockutils [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Lock "c1fc5465-5ac4-4b84-8fa6-f9f4dedbb6c1" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 445.499524] env[61906]: DEBUG oslo_concurrency.lockutils [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Lock "c1fc5465-5ac4-4b84-8fa6-f9f4dedbb6c1" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 445.519358] env[61906]: INFO dbcounter [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Registered counter for database nova_cell0 [ 445.527488] env[61906]: INFO dbcounter [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Registered counter for database nova_cell1 [ 445.530748] env[61906]: DEBUG oslo_db.sqlalchemy.engines [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61906) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 445.531380] env[61906]: DEBUG oslo_db.sqlalchemy.engines [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61906) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 445.535924] env[61906]: ERROR nova.db.main.api [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 445.535924] env[61906]: result = function(*args, **kwargs) [ 445.535924] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 445.535924] env[61906]: return func(*args, **kwargs) [ 445.535924] env[61906]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 445.535924] env[61906]: result = fn(*args, **kwargs) [ 445.535924] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 445.535924] env[61906]: return f(*args, **kwargs) [ 445.535924] env[61906]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 445.535924] env[61906]: return db.service_get_minimum_version(context, binaries) [ 445.535924] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 445.535924] env[61906]: _check_db_access() [ 445.535924] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 445.535924] env[61906]: stacktrace = ''.join(traceback.format_stack()) [ 445.535924] env[61906]: [ 445.537036] env[61906]: ERROR nova.db.main.api [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 445.537036] env[61906]: result = function(*args, **kwargs) [ 445.537036] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 445.537036] env[61906]: return func(*args, **kwargs) [ 445.537036] env[61906]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 445.537036] env[61906]: result = fn(*args, **kwargs) [ 445.537036] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 445.537036] env[61906]: return f(*args, **kwargs) [ 445.537036] env[61906]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 445.537036] env[61906]: return db.service_get_minimum_version(context, binaries) [ 445.537036] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 445.537036] env[61906]: _check_db_access() [ 445.537036] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 445.537036] env[61906]: stacktrace = ''.join(traceback.format_stack()) [ 445.537036] env[61906]: [ 445.537866] env[61906]: WARNING nova.objects.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 445.537866] env[61906]: WARNING nova.objects.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Failed to get minimum service version for cell c1fc5465-5ac4-4b84-8fa6-f9f4dedbb6c1 [ 445.537984] env[61906]: DEBUG oslo_concurrency.lockutils [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Acquiring lock "singleton_lock" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 445.538136] env[61906]: DEBUG oslo_concurrency.lockutils [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Acquired lock "singleton_lock" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 445.538376] env[61906]: DEBUG oslo_concurrency.lockutils [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Releasing lock "singleton_lock" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 445.538694] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Full set of CONF: {{(pid=61906) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 445.538835] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ******************************************************************************** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 445.538959] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Configuration options gathered from: {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 445.539103] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 445.539294] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 445.539422] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ================================================================================ {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 445.539623] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] allow_resize_to_same_host = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.539786] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] arq_binding_timeout = 300 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.539914] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] backdoor_port = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.540053] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] backdoor_socket = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.540219] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] block_device_allocate_retries = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.540379] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] block_device_allocate_retries_interval = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.540545] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cert = self.pem {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.540708] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.540873] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute_monitors = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.541069] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] config_dir = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.541262] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] config_drive_format = iso9660 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.541395] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.541556] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] config_source = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.541721] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] console_host = devstack {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.541882] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] control_exchange = nova {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.542048] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cpu_allocation_ratio = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.542213] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] daemon = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.542389] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] debug = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.542549] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] default_access_ip_network_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.542712] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] default_availability_zone = nova {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.542866] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] default_ephemeral_format = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.543030] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] default_green_pool_size = 1000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.543290] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.543459] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] default_schedule_zone = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.543623] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] disallowed_reset_states = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.543781] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] disk_allocation_ratio = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.543939] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] enable_new_services = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.544134] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] enabled_apis = ['osapi_compute'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.544298] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] enabled_ssl_apis = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.544456] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] flat_injected = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.544608] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] force_config_drive = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.544764] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] force_raw_images = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.544928] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] graceful_shutdown_timeout = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.545099] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] heal_instance_info_cache_interval = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.545316] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] host = cpu-1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.545493] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.545654] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] initial_disk_allocation_ratio = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.545811] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] initial_ram_allocation_ratio = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.546030] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.546200] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] instance_build_timeout = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.546360] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] instance_delete_interval = 300 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.546527] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] instance_format = [instance: %(uuid)s] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.546692] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] instance_name_template = instance-%08x {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.546851] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] instance_usage_audit = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.547029] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] instance_usage_audit_period = month {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.547200] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.547368] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] instances_path = /opt/stack/data/nova/instances {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.547535] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] internal_service_availability_zone = internal {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.547692] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] key = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.547853] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] live_migration_retry_count = 30 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.548027] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] log_color = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.548192] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] log_config_append = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.548358] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.548515] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] log_dir = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.548672] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] log_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.548796] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] log_options = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.548954] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] log_rotate_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.549146] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] log_rotate_interval_type = days {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.549371] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] log_rotation_type = none {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.549504] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.549630] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.549798] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.549960] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.550130] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.550263] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] long_rpc_timeout = 1800 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.550424] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] max_concurrent_builds = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.550576] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] max_concurrent_live_migrations = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.550727] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] max_concurrent_snapshots = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.550880] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] max_local_block_devices = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.551048] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] max_logfile_count = 30 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.551207] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] max_logfile_size_mb = 200 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.551362] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] maximum_instance_delete_attempts = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.551526] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] metadata_listen = 0.0.0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.551687] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] metadata_listen_port = 8775 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.551850] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] metadata_workers = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.552011] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] migrate_max_retries = -1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.552181] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] mkisofs_cmd = genisoimage {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.552387] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] my_block_storage_ip = 10.180.1.21 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.552523] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] my_ip = 10.180.1.21 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.552680] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] network_allocate_retries = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.552853] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.553020] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] osapi_compute_listen = 0.0.0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.553213] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] osapi_compute_listen_port = 8774 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.553392] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] osapi_compute_unique_server_name_scope = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.553562] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] osapi_compute_workers = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.553722] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] password_length = 12 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.553879] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] periodic_enable = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.554051] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] periodic_fuzzy_delay = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.554224] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] pointer_model = usbtablet {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.554388] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] preallocate_images = none {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.554542] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] publish_errors = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.554670] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] pybasedir = /opt/stack/nova {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.554822] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ram_allocation_ratio = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.554975] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] rate_limit_burst = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.555154] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] rate_limit_except_level = CRITICAL {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.555316] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] rate_limit_interval = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.555483] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] reboot_timeout = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.555700] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] reclaim_instance_interval = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.555869] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] record = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.556048] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] reimage_timeout_per_gb = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.556220] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] report_interval = 120 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.556381] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] rescue_timeout = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.556540] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] reserved_host_cpus = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.556696] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] reserved_host_disk_mb = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.556851] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] reserved_host_memory_mb = 512 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.557023] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] reserved_huge_pages = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.557178] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] resize_confirm_window = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.557340] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] resize_fs_using_block_device = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.557494] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] resume_guests_state_on_host_boot = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.557657] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.557818] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] rpc_response_timeout = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.557982] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] run_external_periodic_tasks = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.558170] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] running_deleted_instance_action = reap {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.558331] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] running_deleted_instance_poll_interval = 1800 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.558486] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] running_deleted_instance_timeout = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.558639] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] scheduler_instance_sync_interval = 120 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.558800] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] service_down_time = 720 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.558964] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] servicegroup_driver = db {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.559129] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] shell_completion = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.559288] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] shelved_offload_time = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.559443] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] shelved_poll_interval = 3600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.559605] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] shutdown_timeout = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.559759] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] source_is_ipv6 = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.559914] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ssl_only = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.560167] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.560334] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] sync_power_state_interval = 600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.560490] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] sync_power_state_pool_size = 1000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.560651] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] syslog_log_facility = LOG_USER {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.560805] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] tempdir = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.560960] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] timeout_nbd = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.561170] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] transport_url = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.561344] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] update_resources_interval = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.561504] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] use_cow_images = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.561662] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] use_eventlog = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.561818] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] use_journal = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.561973] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] use_json = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.562144] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] use_rootwrap_daemon = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.562303] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] use_stderr = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.562459] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] use_syslog = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.562615] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vcpu_pin_set = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.562778] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plugging_is_fatal = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.562938] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plugging_timeout = 300 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.563111] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] virt_mkfs = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.563296] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] volume_usage_poll_interval = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.563458] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] watch_log_file = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.563622] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] web = /usr/share/spice-html5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 445.563804] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.563967] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.564144] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.564312] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_concurrency.disable_process_locking = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.564865] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.565065] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.565236] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.565410] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.565580] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.565745] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.565927] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.auth_strategy = keystone {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.566109] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.compute_link_prefix = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.566289] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.566461] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.dhcp_domain = novalocal {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.566626] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.enable_instance_password = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.566788] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.glance_link_prefix = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.566948] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.567131] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.567297] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.instance_list_per_project_cells = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.567454] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.list_records_by_skipping_down_cells = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.567611] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.local_metadata_per_cell = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.567776] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.max_limit = 1000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.567942] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.metadata_cache_expiration = 15 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.568131] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.neutron_default_tenant_id = default {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.568308] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.response_validation = warn {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.568475] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.use_neutron_default_nets = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.568642] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.568803] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.568968] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.569152] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.569325] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.vendordata_dynamic_targets = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.569488] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.vendordata_jsonfile_path = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.569667] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.569857] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.backend = dogpile.cache.memcached {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.570031] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.backend_argument = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.570205] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.config_prefix = cache.oslo {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.570375] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.dead_timeout = 60.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.570536] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.debug_cache_backend = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.570696] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.enable_retry_client = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.570855] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.enable_socket_keepalive = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.571039] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.enabled = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.571235] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.enforce_fips_mode = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.571407] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.expiration_time = 600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.571569] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.hashclient_retry_attempts = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.571736] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.hashclient_retry_delay = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.571898] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.memcache_dead_retry = 300 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.572070] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.memcache_password = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.572239] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.572404] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.572567] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.memcache_pool_maxsize = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.572728] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.572888] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.memcache_sasl_enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.573076] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.573301] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.memcache_socket_timeout = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.573462] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.memcache_username = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.573629] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.proxies = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.573794] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.redis_db = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.573956] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.redis_password = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.574144] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.redis_sentinel_service_name = mymaster {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.574324] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.574493] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.redis_server = localhost:6379 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.574655] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.redis_socket_timeout = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.574815] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.redis_username = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.574976] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.retry_attempts = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.575153] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.retry_delay = 0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.575317] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.socket_keepalive_count = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.575479] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.socket_keepalive_idle = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.575640] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.socket_keepalive_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.575798] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.tls_allowed_ciphers = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.575958] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.tls_cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.576126] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.tls_certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.576377] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.tls_enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.576458] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cache.tls_keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.576614] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.576787] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.auth_type = password {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.576949] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.577139] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.catalog_info = volumev3::publicURL {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.577303] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.577466] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.577626] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.cross_az_attach = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.577785] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.debug = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.577943] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.endpoint_template = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.578115] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.http_retries = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.578278] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.578437] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.578607] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.os_region_name = RegionOne {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.578767] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.578923] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cinder.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.579103] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.579264] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute.cpu_dedicated_set = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.579423] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute.cpu_shared_set = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.579585] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute.image_type_exclude_list = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.579746] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.579906] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute.max_concurrent_disk_ops = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.580076] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute.max_disk_devices_to_attach = -1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.580240] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.580411] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.580572] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute.resource_provider_association_refresh = 300 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.580731] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.580889] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute.shutdown_retry_interval = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.581143] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.581308] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] conductor.workers = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.581490] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] console.allowed_origins = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.581660] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] console.ssl_ciphers = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.581819] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] console.ssl_minimum_version = default {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.581997] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] consoleauth.enforce_session_timeout = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.582166] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] consoleauth.token_ttl = 600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.582341] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.582498] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.582660] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.582819] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.582978] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.583174] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.583340] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.583499] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.583658] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.583812] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.583970] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.region_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.584148] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.584303] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.584469] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.service_type = accelerator {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.584627] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.584782] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.584936] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.585102] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.585281] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.585444] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] cyborg.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.585621] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.backend = sqlalchemy {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.585787] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.connection = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.585947] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.connection_debug = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.586126] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.connection_parameters = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.586292] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.connection_recycle_time = 3600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.586453] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.connection_trace = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.586612] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.db_inc_retry_interval = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.586772] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.db_max_retries = 20 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.586933] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.db_max_retry_interval = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.587103] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.db_retry_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.587264] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.max_overflow = 50 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.587426] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.max_pool_size = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.587584] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.max_retries = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.587748] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.587904] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.mysql_wsrep_sync_wait = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.588066] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.pool_timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.588309] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.retry_interval = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.588384] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.slave_connection = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.588543] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.sqlite_synchronous = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.588699] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] database.use_db_reconnect = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.588871] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.backend = sqlalchemy {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.589045] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.connection = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.589210] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.connection_debug = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.589404] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.connection_parameters = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.589615] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.connection_recycle_time = 3600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.589786] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.connection_trace = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.589947] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.db_inc_retry_interval = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.590125] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.db_max_retries = 20 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.590291] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.db_max_retry_interval = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.590450] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.db_retry_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.590609] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.max_overflow = 50 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.590769] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.max_pool_size = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.590929] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.max_retries = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.591136] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.591305] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.591464] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.pool_timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.591626] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.retry_interval = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.591785] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.slave_connection = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.591944] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] api_database.sqlite_synchronous = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.592134] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] devices.enabled_mdev_types = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.592317] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.592488] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ephemeral_storage_encryption.default_format = luks {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.592649] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ephemeral_storage_encryption.enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.592811] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.592980] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.api_servers = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.593174] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.593350] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.593515] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.593674] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.593832] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.593993] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.debug = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.594200] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.default_trusted_certificate_ids = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.594369] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.enable_certificate_validation = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.594530] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.enable_rbd_download = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.594687] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.594862] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.595030] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.595196] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.595354] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.595522] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.num_retries = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.595686] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.rbd_ceph_conf = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.595878] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.rbd_connect_timeout = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.596065] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.rbd_pool = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.596238] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.rbd_user = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.596399] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.region_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.596559] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.596714] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.596879] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.service_type = image {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.597057] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.597219] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.597377] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.597533] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.597710] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.597874] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.verify_glance_signatures = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.598040] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] glance.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.598260] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] guestfs.debug = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.598431] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] mks.enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.598788] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.598977] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] image_cache.manager_interval = 2400 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.599160] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] image_cache.precache_concurrency = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.599332] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] image_cache.remove_unused_base_images = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.599501] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.599667] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.599844] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] image_cache.subdirectory_name = _base {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.600030] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.api_max_retries = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.600201] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.api_retry_interval = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.600380] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.600570] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.auth_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.600736] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.600894] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.601088] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.601272] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.conductor_group = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.601435] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.601596] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.601754] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.601915] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.602084] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.602246] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.602404] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.602570] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.peer_list = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.602729] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.region_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.602885] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.603057] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.serial_console_state_timeout = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.603247] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.603430] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.service_type = baremetal {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.603590] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.shard = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.603752] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.603910] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.604077] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.604293] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.604493] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.604657] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ironic.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.604840] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.605022] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] key_manager.fixed_key = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.605209] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.605374] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.barbican_api_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.605540] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.barbican_endpoint = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.605707] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.barbican_endpoint_type = public {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.605865] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.barbican_region_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.606047] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.606215] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.606377] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.606536] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.606692] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.606853] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.number_of_retries = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.607021] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.retry_delay = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.607187] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.send_service_user_token = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.607347] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.607503] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.607660] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.verify_ssl = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.607814] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican.verify_ssl_path = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.607976] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican_service_user.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.608147] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican_service_user.auth_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.608303] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican_service_user.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.608457] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican_service_user.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.608615] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican_service_user.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.608770] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican_service_user.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.608924] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican_service_user.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.609096] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican_service_user.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.609252] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] barbican_service_user.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.609490] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vault.approle_role_id = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.609576] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vault.approle_secret_id = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.609739] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vault.kv_mountpoint = secret {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.609894] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vault.kv_path = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.610064] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vault.kv_version = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.610224] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vault.namespace = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.610380] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vault.root_token_id = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.610533] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vault.ssl_ca_crt_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.610694] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vault.timeout = 60.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.610850] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vault.use_ssl = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.611028] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.611230] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.611402] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.auth_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.611620] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.611795] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.611986] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.612133] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.612297] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.612564] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.612674] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.612832] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.612989] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.613189] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.613363] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.region_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.613527] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.613687] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.613855] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.service_type = identity {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.614023] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.614186] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.614357] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.614517] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.614697] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.614856] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] keystone.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.615065] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.connection_uri = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.615229] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.cpu_mode = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.615397] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.cpu_model_extra_flags = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.615564] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.cpu_models = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.615732] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.cpu_power_governor_high = performance {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.615898] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.cpu_power_governor_low = powersave {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.616069] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.cpu_power_management = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.616243] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.616405] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.device_detach_attempts = 8 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.616567] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.device_detach_timeout = 20 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.616732] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.disk_cachemodes = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.616889] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.disk_prefix = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.617060] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.enabled_perf_events = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.617226] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.file_backed_memory = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.617390] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.gid_maps = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.617545] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.hw_disk_discard = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.617705] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.hw_machine_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.617876] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.images_rbd_ceph_conf = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.618048] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.618213] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.618380] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.images_rbd_glance_store_name = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.618545] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.images_rbd_pool = rbd {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.618711] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.images_type = default {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.618867] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.images_volume_group = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.619038] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.inject_key = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.619203] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.inject_partition = -2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.619363] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.inject_password = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.619524] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.iscsi_iface = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.619682] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.iser_use_multipath = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.619843] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.live_migration_bandwidth = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.620008] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.620177] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.live_migration_downtime = 500 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.620339] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.620495] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.620654] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.live_migration_inbound_addr = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.620820] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.620981] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.live_migration_permit_post_copy = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.621154] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.live_migration_scheme = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.621328] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.live_migration_timeout_action = abort {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.621501] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.live_migration_tunnelled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.621661] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.live_migration_uri = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.621824] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.live_migration_with_native_tls = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.621979] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.max_queues = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.622153] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.622405] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.622550] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.nfs_mount_options = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.622842] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.623023] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.623204] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.num_iser_scan_tries = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.623361] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.num_memory_encrypted_guests = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.623525] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.623686] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.num_pcie_ports = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.623849] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.num_volume_scan_tries = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.624024] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.pmem_namespaces = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.624230] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.quobyte_client_cfg = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.624542] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.624717] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.rbd_connect_timeout = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.624882] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.625065] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.625232] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.rbd_secret_uuid = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.625391] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.rbd_user = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.625553] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.625721] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.remote_filesystem_transport = ssh {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.625879] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.rescue_image_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.626045] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.rescue_kernel_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.626204] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.rescue_ramdisk_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.626371] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.626525] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.rx_queue_size = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.626689] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.smbfs_mount_options = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.626967] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.627151] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.snapshot_compression = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.627316] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.snapshot_image_format = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.627532] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.627696] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.sparse_logical_volumes = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.627857] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.swtpm_enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.628033] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.swtpm_group = tss {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.628205] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.swtpm_user = tss {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.628373] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.sysinfo_serial = unique {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.628531] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.tb_cache_size = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.628686] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.tx_queue_size = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.628847] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.uid_maps = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.629012] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.use_virtio_for_bridges = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.629193] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.virt_type = kvm {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.629362] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.volume_clear = zero {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.629523] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.volume_clear_size = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.629686] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.volume_use_multipath = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.629844] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.vzstorage_cache_path = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.630015] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.630190] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.vzstorage_mount_group = qemu {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.630356] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.vzstorage_mount_opts = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.630522] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.630795] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.630969] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.vzstorage_mount_user = stack {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.631152] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.631326] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.631497] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.auth_type = password {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.631655] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.631813] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.631974] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.632151] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.632309] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.632526] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.default_floating_pool = public {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.632627] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.632786] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.extension_sync_interval = 600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.632945] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.http_retries = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.633120] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.633310] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.633472] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.633639] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.633796] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.633964] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.ovs_bridge = br-int {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.634163] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.physnets = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.634349] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.region_name = RegionOne {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.634511] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.634678] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.service_metadata_proxy = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.634838] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.635011] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.service_type = network {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.635182] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.635343] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.635499] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.635654] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.635830] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.635988] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] neutron.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.636176] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] notifications.bdms_in_notifications = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.636401] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] notifications.default_level = INFO {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.636581] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] notifications.notification_format = unversioned {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.636743] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] notifications.notify_on_state_change = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.636917] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.637107] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] pci.alias = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.637284] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] pci.device_spec = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.637451] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] pci.report_in_placement = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.637621] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.637788] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.auth_type = password {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.637954] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.638125] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.638282] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.638444] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.638599] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.638752] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.638907] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.default_domain_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.639070] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.default_domain_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.639229] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.domain_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.639386] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.domain_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.639541] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.639697] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.639853] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.640029] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.640180] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.640347] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.password = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.640503] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.project_domain_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.640670] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.project_domain_name = Default {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.640833] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.project_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.641014] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.project_name = service {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.641217] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.region_name = RegionOne {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.641386] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.641546] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.641713] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.service_type = placement {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.641874] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.642042] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.642204] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.642374] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.system_scope = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.642513] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.642707] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.trust_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.642881] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.user_domain_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.643060] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.user_domain_name = Default {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.643258] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.user_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.643434] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.username = nova {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.643616] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.643776] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] placement.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.643953] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] quota.cores = 20 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.644134] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] quota.count_usage_from_placement = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.644308] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.644480] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] quota.injected_file_content_bytes = 10240 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.644643] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] quota.injected_file_path_length = 255 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.644806] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] quota.injected_files = 5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.644971] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] quota.instances = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.645151] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] quota.key_pairs = 100 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.645318] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] quota.metadata_items = 128 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.645482] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] quota.ram = 51200 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.645643] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] quota.recheck_quota = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.645806] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] quota.server_group_members = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.645970] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] quota.server_groups = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.646152] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.646317] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.646477] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] scheduler.image_metadata_prefilter = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.646634] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.646793] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] scheduler.max_attempts = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.646953] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] scheduler.max_placement_results = 1000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.647125] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.647290] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] scheduler.query_placement_for_image_type_support = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.647445] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.647615] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] scheduler.workers = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.647786] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.647954] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.648144] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.648350] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.648532] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.648697] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.648864] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.649062] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.649236] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.host_subset_size = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.649401] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.649559] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.649718] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.649882] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.isolated_hosts = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.650056] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.isolated_images = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.650223] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.650385] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.650554] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.650716] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.pci_in_placement = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.650875] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.651061] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.651246] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.651412] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.651575] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.651734] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.651893] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.track_instance_changes = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.652082] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.652257] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] metrics.required = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.652478] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] metrics.weight_multiplier = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.652586] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.652751] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] metrics.weight_setting = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.653067] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.653280] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] serial_console.enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.653464] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] serial_console.port_range = 10000:20000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.653637] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.653805] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.653972] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] serial_console.serialproxy_port = 6083 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.654157] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] service_user.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.654331] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] service_user.auth_type = password {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.654492] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] service_user.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.654647] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] service_user.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.654808] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] service_user.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.654968] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] service_user.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.655140] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] service_user.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.655314] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] service_user.send_service_user_token = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.655498] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] service_user.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.655670] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] service_user.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.655842] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] spice.agent_enabled = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.656008] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] spice.enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.656331] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.656524] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.656694] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] spice.html5proxy_port = 6082 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.656858] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] spice.image_compression = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.657024] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] spice.jpeg_compression = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.657189] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] spice.playback_compression = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.657354] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] spice.require_secure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.657523] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] spice.server_listen = 127.0.0.1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.657690] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.657847] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] spice.streaming_mode = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.658013] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] spice.zlib_compression = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.658183] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] upgrade_levels.baseapi = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.658355] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] upgrade_levels.compute = auto {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.658513] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] upgrade_levels.conductor = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.658669] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] upgrade_levels.scheduler = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.658830] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vendordata_dynamic_auth.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.658990] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vendordata_dynamic_auth.auth_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.659163] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vendordata_dynamic_auth.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.659322] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vendordata_dynamic_auth.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.659478] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.659635] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vendordata_dynamic_auth.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.659790] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vendordata_dynamic_auth.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.659948] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.660113] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vendordata_dynamic_auth.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.660290] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.api_retry_count = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.660451] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.ca_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.660621] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.cache_prefix = devstack-image-cache {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.660786] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.cluster_name = testcl1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.660949] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.connection_pool_size = 10 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.661137] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.console_delay_seconds = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.661314] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.datastore_regex = ^datastore.* {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.661519] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.661692] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.host_password = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.661860] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.host_port = 443 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.662038] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.host_username = administrator@vsphere.local {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.662213] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.insecure = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.662377] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.integration_bridge = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.662589] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.maximum_objects = 100 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.662697] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.pbm_default_policy = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.662862] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.pbm_enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.663015] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.pbm_wsdl_location = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.663227] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.663396] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.serial_port_proxy_uri = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.663555] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.serial_port_service_uri = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.663722] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.task_poll_interval = 0.5 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.663892] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.use_linked_clone = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.664071] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.vnc_keymap = en-us {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.664283] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.vnc_port = 5900 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.664453] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vmware.vnc_port_total = 10000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.664640] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vnc.auth_schemes = ['none'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.664817] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vnc.enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.665129] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.665319] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.665492] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vnc.novncproxy_port = 6080 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.665666] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vnc.server_listen = 127.0.0.1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.665836] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.665996] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vnc.vencrypt_ca_certs = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.666171] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vnc.vencrypt_client_cert = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.666330] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vnc.vencrypt_client_key = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.666503] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.666663] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.disable_deep_image_inspection = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.666820] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.666979] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.667152] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.667315] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.disable_rootwrap = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.667476] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.enable_numa_live_migration = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.667635] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.667793] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.667951] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.668122] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.libvirt_disable_apic = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.668283] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.668443] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.668603] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.668763] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.668922] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.669091] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.669328] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.669418] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.669575] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.669738] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.669921] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.670101] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] wsgi.client_socket_timeout = 900 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.670274] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] wsgi.default_pool_size = 1000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.670440] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] wsgi.keep_alive = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.670602] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] wsgi.max_header_line = 16384 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.670763] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] wsgi.secure_proxy_ssl_header = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.670920] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] wsgi.ssl_ca_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.671099] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] wsgi.ssl_cert_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.671267] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] wsgi.ssl_key_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.671432] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] wsgi.tcp_keepidle = 600 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.671610] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.671777] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] zvm.ca_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.671935] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] zvm.cloud_connector_url = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.672232] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.672430] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] zvm.reachable_timeout = 300 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.672589] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_policy.enforce_new_defaults = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.672969] env[61906]: WARNING oslo_config.cfg [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 445.673179] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_policy.enforce_scope = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.673373] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_policy.policy_default_rule = default {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.673559] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.673730] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_policy.policy_file = policy.yaml {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.673898] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.674069] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.674292] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.674473] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.674635] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.674801] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.674976] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.675167] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler.connection_string = messaging:// {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.675335] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler.enabled = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.675503] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler.es_doc_type = notification {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.675663] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler.es_scroll_size = 10000 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.675830] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler.es_scroll_time = 2m {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.675990] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler.filter_error_trace = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.676169] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler.hmac_keys = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.676336] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler.sentinel_service_name = mymaster {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.676499] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler.socket_timeout = 0.1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.676659] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler.trace_requests = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.676816] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler.trace_sqlalchemy = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.676996] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler_jaeger.process_tags = {} {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.677169] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler_jaeger.service_name_prefix = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.677333] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] profiler_otlp.service_name_prefix = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.677496] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] remote_debug.host = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.677656] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] remote_debug.port = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.677832] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.677993] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.678167] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.678355] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.678564] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.678733] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.678894] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.679069] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.679232] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.679403] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.679561] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.679729] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.679895] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.680072] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.680245] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.680411] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.680571] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.680742] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.680902] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.681071] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.681242] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.681406] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.681567] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.681731] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.682184] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.682184] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.682248] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.682361] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.682526] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.682687] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.ssl = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.682855] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.683073] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.683217] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.683393] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.683563] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.ssl_version = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.683723] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.683908] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.684085] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_notifications.retry = -1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.684312] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.684500] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_messaging_notifications.transport_url = **** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.684673] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.auth_section = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.684836] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.auth_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.684994] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.cafile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.685166] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.certfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.685330] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.collect_timing = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.685487] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.connect_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.685643] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.connect_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.685799] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.endpoint_id = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.685956] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.endpoint_override = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.686126] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.insecure = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.686284] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.keyfile = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.686440] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.max_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.686596] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.min_version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.686752] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.region_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.686911] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.retriable_status_codes = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.687079] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.service_name = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.687241] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.service_type = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.687402] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.split_loggers = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.687558] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.status_code_retries = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.687713] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.status_code_retry_delay = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.687868] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.timeout = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.688033] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.valid_interfaces = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.688196] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_limit.version = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.688362] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_reports.file_event_handler = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.688523] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.688681] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] oslo_reports.log_dir = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.688851] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.689015] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.689179] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.689346] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.689507] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.689663] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.689829] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.689988] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plug_ovs_privileged.group = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.690158] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.690325] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.690484] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.690640] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] vif_plug_ovs_privileged.user = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.690810] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_linux_bridge.flat_interface = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.690985] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.691181] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.691342] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.691511] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.691674] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.691837] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.691997] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.692185] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.692355] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_ovs.isolate_vif = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.692521] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.692685] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.692849] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.693028] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_ovs.ovsdb_interface = native {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.693216] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] os_vif_ovs.per_port_bridge = False {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.693394] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] privsep_osbrick.capabilities = [21] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.693553] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] privsep_osbrick.group = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.693710] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] privsep_osbrick.helper_command = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.693873] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.694050] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.694240] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] privsep_osbrick.user = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.694416] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.694576] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] nova_sys_admin.group = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.694733] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] nova_sys_admin.helper_command = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.694896] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.695074] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.695239] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] nova_sys_admin.user = None {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 445.695371] env[61906]: DEBUG oslo_service.service [None req-67b07d88-fff1-4e71-b8ce-78faa615fb5f None None] ******************************************************************************** {{(pid=61906) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 445.695853] env[61906]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 446.199127] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Getting list of instances from cluster (obj){ [ 446.199127] env[61906]: value = "domain-c8" [ 446.199127] env[61906]: _type = "ClusterComputeResource" [ 446.199127] env[61906]: } {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 446.200245] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c405b5-338c-4899-aad8-9fd8a51f4634 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 446.209152] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Got total of 0 instances {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 446.209771] env[61906]: WARNING nova.virt.vmwareapi.driver [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 446.210129] env[61906]: INFO nova.virt.node [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Generated node identity 6e524f5d-985e-41eb-a7ae-996be3ae1c1a [ 446.210367] env[61906]: INFO nova.virt.node [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Wrote node identity 6e524f5d-985e-41eb-a7ae-996be3ae1c1a to /opt/stack/data/n-cpu-1/compute_id [ 446.713484] env[61906]: WARNING nova.compute.manager [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Compute nodes ['6e524f5d-985e-41eb-a7ae-996be3ae1c1a'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 447.719753] env[61906]: INFO nova.compute.manager [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 448.725708] env[61906]: WARNING nova.compute.manager [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 448.726074] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 448.726226] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 448.726374] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 448.726534] env[61906]: DEBUG nova.compute.resource_tracker [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 448.727470] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a938e823-73dc-4bed-9da2-bed77287b227 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 448.735533] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe09409-262e-47ea-8daf-8ce9db2f4638 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 448.748509] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1601d218-9cf6-4b35-871d-7becc523bbb2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 448.754511] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd0e16a-1d28-4380-914a-ed4ec2633e10 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 448.782960] env[61906]: DEBUG nova.compute.resource_tracker [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181507MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 448.783075] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 448.783296] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 449.285840] env[61906]: WARNING nova.compute.resource_tracker [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] No compute node record for cpu-1:6e524f5d-985e-41eb-a7ae-996be3ae1c1a: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 6e524f5d-985e-41eb-a7ae-996be3ae1c1a could not be found. [ 449.790165] env[61906]: INFO nova.compute.resource_tracker [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a [ 451.298178] env[61906]: DEBUG nova.compute.resource_tracker [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 451.298497] env[61906]: DEBUG nova.compute.resource_tracker [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 451.451181] env[61906]: INFO nova.scheduler.client.report [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] [req-365416d7-b57a-4d8c-8b49-a3f0a5f97645] Created resource provider record via placement API for resource provider with UUID 6e524f5d-985e-41eb-a7ae-996be3ae1c1a and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 451.468553] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1340c4f3-5f2d-44fb-84bf-db1e8bea618c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 451.476514] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd4fc77-16b4-4967-929e-84e069bb2fa3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 451.506370] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b3ab46e-d3e1-4a3c-9de2-4240cfd2f758 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 451.513401] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3179adf-9025-4c4b-a67a-009e05b21f1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 451.525898] env[61906]: DEBUG nova.compute.provider_tree [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 452.062054] env[61906]: DEBUG nova.scheduler.client.report [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Updated inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 452.062299] env[61906]: DEBUG nova.compute.provider_tree [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Updating resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a generation from 0 to 1 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 452.062441] env[61906]: DEBUG nova.compute.provider_tree [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 452.116566] env[61906]: DEBUG nova.compute.provider_tree [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Updating resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a generation from 1 to 2 during operation: update_traits {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 452.620716] env[61906]: DEBUG nova.compute.resource_tracker [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 452.621074] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.838s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 452.621120] env[61906]: DEBUG nova.service [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Creating RPC server for service compute {{(pid=61906) start /opt/stack/nova/nova/service.py:186}} [ 452.634794] env[61906]: DEBUG nova.service [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] Join ServiceGroup membership for this service compute {{(pid=61906) start /opt/stack/nova/nova/service.py:203}} [ 452.634986] env[61906]: DEBUG nova.servicegroup.drivers.db [None req-2efc9899-a98b-4acb-b882-6970d7529382 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61906) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 490.684632] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Acquiring lock "2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 490.685235] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Lock "2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 491.191035] env[61906]: DEBUG nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 491.407739] env[61906]: DEBUG oslo_concurrency.lockutils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquiring lock "85effb81-494c-4bda-a093-6dde523d5f45" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 491.407739] env[61906]: DEBUG oslo_concurrency.lockutils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Lock "85effb81-494c-4bda-a093-6dde523d5f45" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 491.750697] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 491.751481] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 491.754442] env[61906]: INFO nova.compute.claims [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 491.828857] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquiring lock "c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 491.831190] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Lock "c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 491.914227] env[61906]: DEBUG nova.compute.manager [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 492.333960] env[61906]: DEBUG nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 492.447740] env[61906]: DEBUG oslo_concurrency.lockutils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 492.861535] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 492.879393] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010de7e8-b3e8-4815-8930-e7f736317013 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.894237] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936748b1-3af0-40ba-aa31-0fb724523e78 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.905573] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Acquiring lock "a3e97bd0-34e0-45f2-a10b-f554e10ca710" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 492.905825] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Lock "a3e97bd0-34e0-45f2-a10b-f554e10ca710" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 492.938909] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e442056d-775a-4f2a-854c-c1e60649a49b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.955389] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2705c158-a3ba-47d6-b2ef-17f8bbe4e4d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.975122] env[61906]: DEBUG nova.compute.provider_tree [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 493.413020] env[61906]: DEBUG nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 493.479344] env[61906]: DEBUG nova.scheduler.client.report [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 493.950103] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 493.970934] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquiring lock "5f417757-4295-4faa-83d2-1ad2d9c04ad6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 493.971245] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Lock "5f417757-4295-4faa-83d2-1ad2d9c04ad6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 493.983859] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.233s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 493.984468] env[61906]: DEBUG nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 493.987819] env[61906]: DEBUG oslo_concurrency.lockutils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.540s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 493.993421] env[61906]: INFO nova.compute.claims [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 494.475037] env[61906]: DEBUG nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 494.499165] env[61906]: DEBUG nova.compute.utils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 494.501712] env[61906]: DEBUG nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 494.501970] env[61906]: DEBUG nova.network.neutron [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 495.008081] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.008930] env[61906]: DEBUG nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 495.132509] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3516de91-1d67-4e83-8726-a84f9fd6af51 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.143945] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc01140-18d7-4abe-b6f7-0291113ef122 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.194600] env[61906]: DEBUG nova.policy [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1acdad002da84093a4591708997fd42f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f8ea11602775475cae99ca4cfd9817f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 495.196326] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54eb7fb1-d23d-46ce-bac1-80e83802bad0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.205054] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f7a9e3-32c5-431b-ac89-6897e7cc8c50 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.223755] env[61906]: DEBUG nova.compute.provider_tree [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 495.727588] env[61906]: DEBUG nova.scheduler.client.report [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 496.024228] env[61906]: DEBUG nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 496.065305] env[61906]: DEBUG nova.virt.hardware [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 496.065305] env[61906]: DEBUG nova.virt.hardware [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 496.065305] env[61906]: DEBUG nova.virt.hardware [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 496.065305] env[61906]: DEBUG nova.virt.hardware [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 496.065595] env[61906]: DEBUG nova.virt.hardware [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 496.066623] env[61906]: DEBUG nova.virt.hardware [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 496.066623] env[61906]: DEBUG nova.virt.hardware [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 496.066623] env[61906]: DEBUG nova.virt.hardware [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 496.067189] env[61906]: DEBUG nova.virt.hardware [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 496.067476] env[61906]: DEBUG nova.virt.hardware [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 496.068019] env[61906]: DEBUG nova.virt.hardware [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 496.069394] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5019c2b-c5be-4172-b5bc-08bd1c5d6622 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.078872] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a619575c-b1b6-4b85-b379-19d51ea6f8d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.096368] env[61906]: DEBUG nova.network.neutron [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Successfully created port: 7f55ae0f-f69b-416e-9508-5c259ddb945d {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 496.100830] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d82f1ad8-52cd-4270-b40c-fd60a9a5b76a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.233094] env[61906]: DEBUG oslo_concurrency.lockutils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.246s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 496.237522] env[61906]: DEBUG nova.compute.manager [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 496.238229] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.377s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.239940] env[61906]: INFO nova.compute.claims [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 496.746485] env[61906]: DEBUG nova.compute.utils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 496.758883] env[61906]: DEBUG nova.compute.manager [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 497.257058] env[61906]: DEBUG nova.compute.manager [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 497.379280] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29256d6d-d3d6-4353-9703-9f8ee4c8a9cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.388254] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15afc4a2-ff51-4a48-a9a4-ea5e581f7b3c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.421898] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c000779d-9406-4f8a-85f3-9102bb786b03 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.432656] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c527bde7-4d33-4391-9e7a-d332a55ee265 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.445688] env[61906]: DEBUG nova.compute.provider_tree [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 497.639095] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Acquiring lock "17b85f46-4660-4af6-825d-3f2e7ea931ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 497.639319] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Lock "17b85f46-4660-4af6-825d-3f2e7ea931ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 497.701889] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "2f64e954-bf3e-464f-a72e-d83f0783af5c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 497.702314] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "2f64e954-bf3e-464f-a72e-d83f0783af5c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 497.951153] env[61906]: DEBUG nova.scheduler.client.report [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 498.145171] env[61906]: DEBUG nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 498.208235] env[61906]: DEBUG nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 498.266663] env[61906]: DEBUG nova.compute.manager [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 498.311826] env[61906]: DEBUG nova.virt.hardware [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 498.311826] env[61906]: DEBUG nova.virt.hardware [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 498.311826] env[61906]: DEBUG nova.virt.hardware [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 498.311826] env[61906]: DEBUG nova.virt.hardware [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 498.312371] env[61906]: DEBUG nova.virt.hardware [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 498.312371] env[61906]: DEBUG nova.virt.hardware [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 498.312371] env[61906]: DEBUG nova.virt.hardware [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 498.312507] env[61906]: DEBUG nova.virt.hardware [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 498.312905] env[61906]: DEBUG nova.virt.hardware [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 498.313124] env[61906]: DEBUG nova.virt.hardware [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 498.313309] env[61906]: DEBUG nova.virt.hardware [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 498.315185] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4c84ba-e6ed-4833-984e-c692264c3edf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.324105] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812136e4-3ff9-49ba-869b-65e989ea85be {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.343495] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 498.353139] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 498.353879] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ea692920-06bb-4ad3-91f0-9b99d45f9073 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.365225] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Created folder: OpenStack in parent group-v4. [ 498.365733] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Creating folder: Project (a1f80a3d096542b6a0da3203044718ae). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 498.365733] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-de95df14-c5c2-4043-84c0-6a3d70a665cc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.374541] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Created folder: Project (a1f80a3d096542b6a0da3203044718ae) in parent group-v288914. [ 498.374742] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Creating folder: Instances. Parent ref: group-v288915. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 498.375022] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-176180df-04c3-432e-9346-2e6f777811c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.384227] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Created folder: Instances in parent group-v288915. [ 498.384227] env[61906]: DEBUG oslo.service.loopingcall [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 498.384227] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 498.384335] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e552c76-2a51-455d-ade0-1e7580b9c16e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.405565] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 498.405565] env[61906]: value = "task-1356423" [ 498.405565] env[61906]: _type = "Task" [ 498.405565] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 498.418316] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356423, 'name': CreateVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 498.459645] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.222s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 498.460169] env[61906]: DEBUG nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 498.466032] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.514s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 498.466032] env[61906]: INFO nova.compute.claims [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 498.686944] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 498.745155] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 498.921305] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356423, 'name': CreateVM_Task, 'duration_secs': 0.303325} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 498.921493] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 498.922493] env[61906]: DEBUG oslo_vmware.service [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dedec142-c409-4d55-8dd9-cf7573ad3065 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.933876] env[61906]: DEBUG oslo_concurrency.lockutils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 498.934249] env[61906]: DEBUG oslo_concurrency.lockutils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 498.935289] env[61906]: DEBUG oslo_concurrency.lockutils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 498.935592] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c325e602-bc2a-4772-8f2d-ef33bf7f6449 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.941385] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 498.941385] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528d82b5-e60b-5bbb-3623-4e4b9af6f1fe" [ 498.941385] env[61906]: _type = "Task" [ 498.941385] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 498.952477] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528d82b5-e60b-5bbb-3623-4e4b9af6f1fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 498.971610] env[61906]: DEBUG nova.compute.utils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 498.976352] env[61906]: DEBUG nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 498.977789] env[61906]: DEBUG nova.network.neutron [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 499.117171] env[61906]: DEBUG nova.policy [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ced8ec3ae40e4d44a2ad41f74784b172', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ae83ff4f61b47ec884177f50ac9d303', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 499.452935] env[61906]: DEBUG oslo_concurrency.lockutils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 499.453303] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 499.453810] env[61906]: DEBUG oslo_concurrency.lockutils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 499.453810] env[61906]: DEBUG oslo_concurrency.lockutils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 499.454994] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 499.455693] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7a03fb64-0621-409a-ab57-4cdde99db375 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.474238] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 499.474923] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 499.476632] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb5190b7-0b0d-4ad5-93e3-576b642f0595 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.479447] env[61906]: DEBUG nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 499.492019] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8778c03-3f61-4256-a410-4de0625aabae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.499340] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 499.499340] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52741656-3231-2658-db1d-7cde6f6f44e0" [ 499.499340] env[61906]: _type = "Task" [ 499.499340] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 499.515181] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52741656-3231-2658-db1d-7cde6f6f44e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 499.661708] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a99f3500-3f3b-4987-afcd-65fd81cde78b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.671854] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Acquiring lock "7f4bb9fb-fa64-451d-9876-7e4314c13a64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 499.673020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Lock "7f4bb9fb-fa64-451d-9876-7e4314c13a64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 499.682014] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8b38d8-84c7-4e5e-971b-56976d00a511 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.720258] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c61adcf-a805-4a25-a44e-850f36578fa1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.729632] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c46ff7e-ce9f-4548-af90-c187299717b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.746946] env[61906]: DEBUG nova.compute.provider_tree [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 499.877911] env[61906]: ERROR nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7f55ae0f-f69b-416e-9508-5c259ddb945d, please check neutron logs for more information. [ 499.877911] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 499.877911] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 499.877911] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 499.877911] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 499.877911] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 499.877911] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 499.877911] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 499.877911] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 499.877911] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 499.877911] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 499.877911] env[61906]: ERROR nova.compute.manager raise self.value [ 499.877911] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 499.877911] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 499.877911] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 499.877911] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 499.878429] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 499.878429] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 499.878429] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7f55ae0f-f69b-416e-9508-5c259ddb945d, please check neutron logs for more information. [ 499.878429] env[61906]: ERROR nova.compute.manager [ 499.878429] env[61906]: Traceback (most recent call last): [ 499.878429] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 499.878429] env[61906]: listener.cb(fileno) [ 499.878429] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 499.878429] env[61906]: result = function(*args, **kwargs) [ 499.878429] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 499.878429] env[61906]: return func(*args, **kwargs) [ 499.878429] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 499.878429] env[61906]: raise e [ 499.878429] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 499.878429] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 499.878429] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 499.878429] env[61906]: created_port_ids = self._update_ports_for_instance( [ 499.878429] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 499.878429] env[61906]: with excutils.save_and_reraise_exception(): [ 499.878429] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 499.878429] env[61906]: self.force_reraise() [ 499.878429] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 499.878429] env[61906]: raise self.value [ 499.878429] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 499.878429] env[61906]: updated_port = self._update_port( [ 499.878429] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 499.878429] env[61906]: _ensure_no_port_binding_failure(port) [ 499.878429] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 499.878429] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 499.879329] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 7f55ae0f-f69b-416e-9508-5c259ddb945d, please check neutron logs for more information. [ 499.879329] env[61906]: Removing descriptor: 15 [ 499.879787] env[61906]: ERROR nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7f55ae0f-f69b-416e-9508-5c259ddb945d, please check neutron logs for more information. [ 499.879787] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Traceback (most recent call last): [ 499.879787] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 499.879787] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] yield resources [ 499.879787] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 499.879787] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] self.driver.spawn(context, instance, image_meta, [ 499.879787] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 499.879787] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 499.879787] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 499.879787] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] vm_ref = self.build_virtual_machine(instance, [ 499.879787] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 499.880084] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] vif_infos = vmwarevif.get_vif_info(self._session, [ 499.880084] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 499.880084] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] for vif in network_info: [ 499.880084] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 499.880084] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] return self._sync_wrapper(fn, *args, **kwargs) [ 499.880084] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 499.880084] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] self.wait() [ 499.880084] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 499.880084] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] self[:] = self._gt.wait() [ 499.880084] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 499.880084] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] return self._exit_event.wait() [ 499.880084] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 499.880084] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] result = hub.switch() [ 499.880539] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 499.880539] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] return self.greenlet.switch() [ 499.880539] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 499.880539] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] result = function(*args, **kwargs) [ 499.880539] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 499.880539] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] return func(*args, **kwargs) [ 499.880539] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 499.880539] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] raise e [ 499.880539] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 499.880539] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] nwinfo = self.network_api.allocate_for_instance( [ 499.880539] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 499.880539] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] created_port_ids = self._update_ports_for_instance( [ 499.880539] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 499.880947] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] with excutils.save_and_reraise_exception(): [ 499.880947] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 499.880947] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] self.force_reraise() [ 499.880947] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 499.880947] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] raise self.value [ 499.880947] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 499.880947] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] updated_port = self._update_port( [ 499.880947] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 499.880947] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] _ensure_no_port_binding_failure(port) [ 499.880947] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 499.880947] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] raise exception.PortBindingFailed(port_id=port['id']) [ 499.880947] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] nova.exception.PortBindingFailed: Binding failed for port 7f55ae0f-f69b-416e-9508-5c259ddb945d, please check neutron logs for more information. [ 499.880947] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] [ 499.881283] env[61906]: INFO nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Terminating instance [ 499.884377] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Acquiring lock "refresh_cache-2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 499.884634] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Acquired lock "refresh_cache-2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 499.884891] env[61906]: DEBUG nova.network.neutron [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 500.013799] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Preparing fetch location {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 500.014099] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Creating directory with path [datastore2] vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 500.014608] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-37adfaa4-24e9-4a91-abc9-aac22ea26a85 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.021898] env[61906]: DEBUG nova.network.neutron [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Successfully created port: e27c9e5a-bf20-4fde-aa0d-0f7859533c1c {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 500.039624] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Created directory with path [datastore2] vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 500.039624] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Fetch image to [datastore2] vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 500.039852] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Downloading image file data 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 to [datastore2] vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk on the data store datastore2 {{(pid=61906) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 500.040574] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e63f42-dc68-4a53-bcb7-d07c83e41dd2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.048270] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ccf59d3-b11e-427f-b213-18282f3614cc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.059585] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7057b0e7-1017-4a9f-ae19-e0838b7d72e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.094627] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0030d425-8f05-49be-8251-47ff7431d1a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.100980] env[61906]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-2a61683a-896a-4e65-aea2-326c2376c176 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.123753] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Downloading image file data 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 to the data store datastore2 {{(pid=61906) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 500.178157] env[61906]: DEBUG nova.compute.manager [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 500.207418] env[61906]: DEBUG oslo_vmware.rw_handles [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61906) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 500.281314] env[61906]: DEBUG nova.scheduler.client.report [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 500.495280] env[61906]: DEBUG nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 500.531760] env[61906]: DEBUG nova.virt.hardware [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 500.531760] env[61906]: DEBUG nova.virt.hardware [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 500.531760] env[61906]: DEBUG nova.virt.hardware [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 500.532179] env[61906]: DEBUG nova.virt.hardware [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 500.532179] env[61906]: DEBUG nova.virt.hardware [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 500.532179] env[61906]: DEBUG nova.virt.hardware [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 500.532179] env[61906]: DEBUG nova.virt.hardware [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 500.532179] env[61906]: DEBUG nova.virt.hardware [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 500.532464] env[61906]: DEBUG nova.virt.hardware [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 500.532464] env[61906]: DEBUG nova.virt.hardware [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 500.532464] env[61906]: DEBUG nova.virt.hardware [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 500.532552] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc38bfb-e08c-4cb8-9b9a-d5cc3ccc49f3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.549189] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62129f99-cc4f-4baf-9613-cb429ee240c5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.599091] env[61906]: DEBUG nova.network.neutron [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 500.717013] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 500.788175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.325s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 500.788710] env[61906]: DEBUG nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 500.792751] env[61906]: DEBUG nova.compute.manager [req-5c2ae6c6-ed98-4673-8209-1d264d78c42b req-6f74a748-e5e6-49ed-b168-94ce712c59d2 service nova] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Received event network-changed-7f55ae0f-f69b-416e-9508-5c259ddb945d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 500.792992] env[61906]: DEBUG nova.compute.manager [req-5c2ae6c6-ed98-4673-8209-1d264d78c42b req-6f74a748-e5e6-49ed-b168-94ce712c59d2 service nova] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Refreshing instance network info cache due to event network-changed-7f55ae0f-f69b-416e-9508-5c259ddb945d. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 500.793939] env[61906]: DEBUG oslo_concurrency.lockutils [req-5c2ae6c6-ed98-4673-8209-1d264d78c42b req-6f74a748-e5e6-49ed-b168-94ce712c59d2 service nova] Acquiring lock "refresh_cache-2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 500.795970] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.788s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 500.797348] env[61906]: INFO nova.compute.claims [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 500.897308] env[61906]: DEBUG oslo_vmware.rw_handles [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Completed reading data from the image iterator. {{(pid=61906) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 500.897510] env[61906]: DEBUG oslo_vmware.rw_handles [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 501.033391] env[61906]: DEBUG nova.network.neutron [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 501.176704] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Downloaded image file data 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 to vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk on the data store datastore2 {{(pid=61906) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 501.178696] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Caching image {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 501.178949] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Copying Virtual Disk [datastore2] vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk to [datastore2] vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 501.179580] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4609ed2-eded-413c-a287-57574d0c4ad1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.187557] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 501.187557] env[61906]: value = "task-1356424" [ 501.187557] env[61906]: _type = "Task" [ 501.187557] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 501.197588] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356424, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 501.295788] env[61906]: DEBUG nova.compute.utils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 501.300949] env[61906]: DEBUG nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 501.300949] env[61906]: DEBUG nova.network.neutron [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 501.484924] env[61906]: DEBUG nova.policy [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1b6731041d8d42309ece30a6e5501ff5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1043acccd324087a137dd7109047669', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 501.537439] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Releasing lock "refresh_cache-2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 501.538035] env[61906]: DEBUG nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 501.538912] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 501.539308] env[61906]: DEBUG oslo_concurrency.lockutils [req-5c2ae6c6-ed98-4673-8209-1d264d78c42b req-6f74a748-e5e6-49ed-b168-94ce712c59d2 service nova] Acquired lock "refresh_cache-2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 501.539494] env[61906]: DEBUG nova.network.neutron [req-5c2ae6c6-ed98-4673-8209-1d264d78c42b req-6f74a748-e5e6-49ed-b168-94ce712c59d2 service nova] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Refreshing network info cache for port 7f55ae0f-f69b-416e-9508-5c259ddb945d {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 501.540640] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23f4f186-4508-443e-b93b-8f177e288300 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.552035] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde138a7-4d82-4768-9522-c8031859755d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.578846] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3 could not be found. [ 501.579086] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 501.579512] env[61906]: INFO nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 501.579774] env[61906]: DEBUG oslo.service.loopingcall [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 501.580897] env[61906]: DEBUG nova.compute.manager [-] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 501.580897] env[61906]: DEBUG nova.network.neutron [-] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 501.636352] env[61906]: DEBUG nova.network.neutron [-] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 501.701243] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356424, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 501.804578] env[61906]: DEBUG nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 501.987255] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c795e568-7a74-45c4-95e7-f926c6db59c5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.996097] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a97d945-0dde-4fe3-9212-5683ee45c899 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.043200] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1bac048-9540-445c-a89d-7f21ba3d791d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.057770] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd42d92a-5bc4-4d29-a839-3d62b29f9827 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.078305] env[61906]: DEBUG nova.compute.provider_tree [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 502.142174] env[61906]: DEBUG nova.network.neutron [-] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 502.178199] env[61906]: DEBUG nova.network.neutron [req-5c2ae6c6-ed98-4673-8209-1d264d78c42b req-6f74a748-e5e6-49ed-b168-94ce712c59d2 service nova] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 502.208148] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356424, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 502.582719] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 502.646897] env[61906]: INFO nova.compute.manager [-] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Took 1.07 seconds to deallocate network for instance. [ 502.647999] env[61906]: DEBUG nova.compute.claims [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 502.648246] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 502.700757] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356424, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.347768} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 502.701098] env[61906]: DEBUG nova.network.neutron [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Successfully created port: 2c453256-9d51-46a3-b814-54223f5fc641 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 502.703542] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Copied Virtual Disk [datastore2] vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk to [datastore2] vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 502.703771] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Deleting the datastore file [datastore2] vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 502.703918] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3befb89c-0bcb-4745-878b-ff14d70c5dcf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.712179] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 502.712179] env[61906]: value = "task-1356425" [ 502.712179] env[61906]: _type = "Task" [ 502.712179] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 502.721536] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356425, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 502.748927] env[61906]: DEBUG nova.network.neutron [req-5c2ae6c6-ed98-4673-8209-1d264d78c42b req-6f74a748-e5e6-49ed-b168-94ce712c59d2 service nova] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 502.818268] env[61906]: DEBUG nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 502.854690] env[61906]: DEBUG nova.virt.hardware [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 502.855206] env[61906]: DEBUG nova.virt.hardware [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 502.855529] env[61906]: DEBUG nova.virt.hardware [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 502.859768] env[61906]: DEBUG nova.virt.hardware [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 502.859768] env[61906]: DEBUG nova.virt.hardware [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 502.859768] env[61906]: DEBUG nova.virt.hardware [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 502.859768] env[61906]: DEBUG nova.virt.hardware [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 502.859768] env[61906]: DEBUG nova.virt.hardware [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 502.859985] env[61906]: DEBUG nova.virt.hardware [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 502.859985] env[61906]: DEBUG nova.virt.hardware [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 502.859985] env[61906]: DEBUG nova.virt.hardware [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 502.859985] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c76f0fc-ddd7-4db3-9390-171c8abc2e5f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.872771] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04cbc3cf-10b6-4dec-98a6-fb92d825f2c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.093922] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.296s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 503.093922] env[61906]: DEBUG nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 503.095429] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.409s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 503.097211] env[61906]: INFO nova.compute.claims [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 503.227518] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356425, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.021389} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 503.228957] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 503.229276] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Moving file from [datastore2] vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 to [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1. {{(pid=61906) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 503.229886] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-941770f3-60b1-4f8d-b88f-d8fcb2e69f33 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.241718] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 503.241718] env[61906]: value = "task-1356426" [ 503.241718] env[61906]: _type = "Task" [ 503.241718] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 503.251963] env[61906]: DEBUG oslo_concurrency.lockutils [req-5c2ae6c6-ed98-4673-8209-1d264d78c42b req-6f74a748-e5e6-49ed-b168-94ce712c59d2 service nova] Releasing lock "refresh_cache-2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 503.254331] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356426, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 503.605477] env[61906]: DEBUG nova.compute.utils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 503.609697] env[61906]: DEBUG nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 503.609882] env[61906]: DEBUG nova.network.neutron [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 503.757170] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356426, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024299} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 503.761587] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] File moved {{(pid=61906) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 503.761587] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Cleaning up location [datastore2] vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 503.761587] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Deleting the datastore file [datastore2] vmware_temp/66992f0c-8bb8-470f-9b65-38712795ed04 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 503.761587] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f81ecee7-0cce-428a-91ad-6211c5a2f7bf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.766155] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 503.766155] env[61906]: value = "task-1356427" [ 503.766155] env[61906]: _type = "Task" [ 503.766155] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 503.775683] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356427, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 503.830149] env[61906]: DEBUG nova.compute.manager [req-1226769c-c9f2-483e-a604-1250009eb322 req-3780b461-ea4b-4fc2-b394-2fe3e48c8306 service nova] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Received event network-vif-deleted-7f55ae0f-f69b-416e-9508-5c259ddb945d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 504.011019] env[61906]: DEBUG nova.policy [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0c0b016b8f1a4c569d8a0e7749ba0a2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2d1343211154305bb672928d6506d94', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 504.114100] env[61906]: DEBUG nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 504.280533] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356427, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025615} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 504.280613] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 504.283679] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b408729-fe50-4e6d-a045-b5e281e37d4c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.291200] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 504.291200] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]523e9e97-a08d-60a7-25c2-33f223e2cc2c" [ 504.291200] env[61906]: _type = "Task" [ 504.291200] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 504.303265] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]523e9e97-a08d-60a7-25c2-33f223e2cc2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 504.407674] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7cea9d6-0638-4d19-9ff5-7a6967d7b8d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.415728] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4bc339-8c6c-4d92-a861-dae5f46879ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.460508] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246e3a5c-51e4-48d2-9906-89384c65e89b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.469731] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103ac4bd-b8d0-4fba-9225-10480cd5a793 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.485163] env[61906]: DEBUG nova.compute.provider_tree [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 504.809914] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]523e9e97-a08d-60a7-25c2-33f223e2cc2c, 'name': SearchDatastore_Task, 'duration_secs': 0.008964} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 504.810241] env[61906]: DEBUG oslo_concurrency.lockutils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 504.810455] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45/85effb81-494c-4bda-a093-6dde523d5f45.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 504.810715] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b647bbcf-b774-426a-8bed-8895782cccd5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.821236] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 504.821236] env[61906]: value = "task-1356428" [ 504.821236] env[61906]: _type = "Task" [ 504.821236] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 504.833138] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356428, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 504.939397] env[61906]: ERROR nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e27c9e5a-bf20-4fde-aa0d-0f7859533c1c, please check neutron logs for more information. [ 504.939397] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 504.939397] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 504.939397] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 504.939397] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 504.939397] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 504.939397] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 504.939397] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 504.939397] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 504.939397] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 504.939397] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 504.939397] env[61906]: ERROR nova.compute.manager raise self.value [ 504.939397] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 504.939397] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 504.939397] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 504.939397] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 504.940150] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 504.940150] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 504.940150] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e27c9e5a-bf20-4fde-aa0d-0f7859533c1c, please check neutron logs for more information. [ 504.940150] env[61906]: ERROR nova.compute.manager [ 504.940150] env[61906]: Traceback (most recent call last): [ 504.940150] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 504.940150] env[61906]: listener.cb(fileno) [ 504.940150] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 504.940150] env[61906]: result = function(*args, **kwargs) [ 504.940150] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 504.940150] env[61906]: return func(*args, **kwargs) [ 504.940150] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 504.940150] env[61906]: raise e [ 504.940150] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 504.940150] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 504.940150] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 504.940150] env[61906]: created_port_ids = self._update_ports_for_instance( [ 504.940150] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 504.940150] env[61906]: with excutils.save_and_reraise_exception(): [ 504.940150] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 504.940150] env[61906]: self.force_reraise() [ 504.940150] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 504.940150] env[61906]: raise self.value [ 504.940150] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 504.940150] env[61906]: updated_port = self._update_port( [ 504.940150] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 504.940150] env[61906]: _ensure_no_port_binding_failure(port) [ 504.940150] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 504.940150] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 504.941679] env[61906]: nova.exception.PortBindingFailed: Binding failed for port e27c9e5a-bf20-4fde-aa0d-0f7859533c1c, please check neutron logs for more information. [ 504.941679] env[61906]: Removing descriptor: 17 [ 504.941679] env[61906]: ERROR nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e27c9e5a-bf20-4fde-aa0d-0f7859533c1c, please check neutron logs for more information. [ 504.941679] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Traceback (most recent call last): [ 504.941679] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 504.941679] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] yield resources [ 504.941679] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 504.941679] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] self.driver.spawn(context, instance, image_meta, [ 504.941679] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 504.941679] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 504.941679] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 504.941679] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] vm_ref = self.build_virtual_machine(instance, [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] for vif in network_info: [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] return self._sync_wrapper(fn, *args, **kwargs) [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] self.wait() [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] self[:] = self._gt.wait() [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] return self._exit_event.wait() [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 504.942411] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] result = hub.switch() [ 504.946457] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 504.946457] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] return self.greenlet.switch() [ 504.946457] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 504.946457] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] result = function(*args, **kwargs) [ 504.946457] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 504.946457] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] return func(*args, **kwargs) [ 504.946457] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 504.946457] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] raise e [ 504.946457] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 504.946457] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] nwinfo = self.network_api.allocate_for_instance( [ 504.946457] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 504.946457] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] created_port_ids = self._update_ports_for_instance( [ 504.946457] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 504.947030] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] with excutils.save_and_reraise_exception(): [ 504.947030] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 504.947030] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] self.force_reraise() [ 504.947030] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 504.947030] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] raise self.value [ 504.947030] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 504.947030] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] updated_port = self._update_port( [ 504.947030] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 504.947030] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] _ensure_no_port_binding_failure(port) [ 504.947030] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 504.947030] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] raise exception.PortBindingFailed(port_id=port['id']) [ 504.947030] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] nova.exception.PortBindingFailed: Binding failed for port e27c9e5a-bf20-4fde-aa0d-0f7859533c1c, please check neutron logs for more information. [ 504.947030] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] [ 504.947422] env[61906]: INFO nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Terminating instance [ 504.947422] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquiring lock "refresh_cache-c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 504.947422] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquired lock "refresh_cache-c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 504.947422] env[61906]: DEBUG nova.network.neutron [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 504.992709] env[61906]: DEBUG nova.scheduler.client.report [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 505.130273] env[61906]: DEBUG nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 505.176028] env[61906]: DEBUG nova.virt.hardware [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 505.176953] env[61906]: DEBUG nova.virt.hardware [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 505.177174] env[61906]: DEBUG nova.virt.hardware [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 505.177463] env[61906]: DEBUG nova.virt.hardware [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 505.177604] env[61906]: DEBUG nova.virt.hardware [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 505.177776] env[61906]: DEBUG nova.virt.hardware [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 505.178051] env[61906]: DEBUG nova.virt.hardware [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 505.178367] env[61906]: DEBUG nova.virt.hardware [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 505.178423] env[61906]: DEBUG nova.virt.hardware [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 505.178560] env[61906]: DEBUG nova.virt.hardware [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 505.178768] env[61906]: DEBUG nova.virt.hardware [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 505.179776] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c9f181-4f7c-4f64-b004-f9d248d664bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.193883] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40086cf7-d3e2-479a-9123-8e39c72010a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.204927] env[61906]: DEBUG nova.compute.manager [req-ea54033b-4fdf-4b00-b1e3-7785982baa0c req-4fc974fa-be5e-4d30-a12a-2d239132e319 service nova] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Received event network-changed-e27c9e5a-bf20-4fde-aa0d-0f7859533c1c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 505.205442] env[61906]: DEBUG nova.compute.manager [req-ea54033b-4fdf-4b00-b1e3-7785982baa0c req-4fc974fa-be5e-4d30-a12a-2d239132e319 service nova] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Refreshing instance network info cache due to event network-changed-e27c9e5a-bf20-4fde-aa0d-0f7859533c1c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 505.205618] env[61906]: DEBUG oslo_concurrency.lockutils [req-ea54033b-4fdf-4b00-b1e3-7785982baa0c req-4fc974fa-be5e-4d30-a12a-2d239132e319 service nova] Acquiring lock "refresh_cache-c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 505.331649] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356428, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504326} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 505.331928] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45/85effb81-494c-4bda-a093-6dde523d5f45.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 505.332642] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 505.332642] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ab831330-8bc1-46e8-9b63-c5a1bce5c43a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.342210] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 505.342210] env[61906]: value = "task-1356429" [ 505.342210] env[61906]: _type = "Task" [ 505.342210] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 505.354044] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356429, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 505.500670] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 505.501238] env[61906]: DEBUG nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 505.507374] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.762s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 505.508892] env[61906]: INFO nova.compute.claims [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 505.515400] env[61906]: DEBUG nova.network.neutron [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 505.671383] env[61906]: DEBUG nova.network.neutron [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Successfully created port: 5fbe7162-22fd-463d-a62c-1725806badc1 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 505.783159] env[61906]: DEBUG nova.network.neutron [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 505.854491] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356429, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.156359} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 505.854491] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 505.855422] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec0623ec-4374-4b1c-916c-efdca3a6240f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.882388] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45/85effb81-494c-4bda-a093-6dde523d5f45.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 505.882975] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3627dbd-2fb9-489b-aad8-70b6dc83ba23 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.906701] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 505.906701] env[61906]: value = "task-1356430" [ 505.906701] env[61906]: _type = "Task" [ 505.906701] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 505.916480] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356430, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 506.014365] env[61906]: DEBUG nova.compute.utils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 506.019138] env[61906]: DEBUG nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 506.019338] env[61906]: DEBUG nova.network.neutron [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 506.195737] env[61906]: DEBUG nova.policy [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '51cce259592e406f8d9fc768946d1133', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e40d7235c3f4f6093e5c1cb2cea28ff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 506.286241] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Releasing lock "refresh_cache-c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 506.286241] env[61906]: DEBUG nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 506.286241] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 506.286241] env[61906]: DEBUG oslo_concurrency.lockutils [req-ea54033b-4fdf-4b00-b1e3-7785982baa0c req-4fc974fa-be5e-4d30-a12a-2d239132e319 service nova] Acquired lock "refresh_cache-c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 506.286241] env[61906]: DEBUG nova.network.neutron [req-ea54033b-4fdf-4b00-b1e3-7785982baa0c req-4fc974fa-be5e-4d30-a12a-2d239132e319 service nova] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Refreshing network info cache for port e27c9e5a-bf20-4fde-aa0d-0f7859533c1c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 506.286458] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-365c1c6a-5445-40a1-8151-213b61b19612 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.303739] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985f5439-2576-401b-a94a-461ca97edfbc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.333971] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1 could not be found. [ 506.334209] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 506.334393] env[61906]: INFO nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Took 0.05 seconds to destroy the instance on the hypervisor. [ 506.334637] env[61906]: DEBUG oslo.service.loopingcall [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 506.334839] env[61906]: DEBUG nova.compute.manager [-] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 506.334927] env[61906]: DEBUG nova.network.neutron [-] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 506.388714] env[61906]: DEBUG nova.network.neutron [-] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 506.421290] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356430, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 506.521631] env[61906]: DEBUG nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 506.549968] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Acquiring lock "fa0cbe6c-937b-4276-a419-0ab6e92e795f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.550381] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Lock "fa0cbe6c-937b-4276-a419-0ab6e92e795f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 506.686143] env[61906]: ERROR nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2c453256-9d51-46a3-b814-54223f5fc641, please check neutron logs for more information. [ 506.686143] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 506.686143] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 506.686143] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 506.686143] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 506.686143] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 506.686143] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 506.686143] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 506.686143] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 506.686143] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 506.686143] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 506.686143] env[61906]: ERROR nova.compute.manager raise self.value [ 506.686143] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 506.686143] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 506.686143] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 506.686143] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 506.686883] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 506.686883] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 506.686883] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2c453256-9d51-46a3-b814-54223f5fc641, please check neutron logs for more information. [ 506.686883] env[61906]: ERROR nova.compute.manager [ 506.686883] env[61906]: Traceback (most recent call last): [ 506.686883] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 506.686883] env[61906]: listener.cb(fileno) [ 506.686883] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 506.686883] env[61906]: result = function(*args, **kwargs) [ 506.686883] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 506.686883] env[61906]: return func(*args, **kwargs) [ 506.686883] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 506.686883] env[61906]: raise e [ 506.686883] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 506.686883] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 506.686883] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 506.686883] env[61906]: created_port_ids = self._update_ports_for_instance( [ 506.686883] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 506.686883] env[61906]: with excutils.save_and_reraise_exception(): [ 506.686883] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 506.686883] env[61906]: self.force_reraise() [ 506.686883] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 506.686883] env[61906]: raise self.value [ 506.686883] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 506.686883] env[61906]: updated_port = self._update_port( [ 506.686883] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 506.686883] env[61906]: _ensure_no_port_binding_failure(port) [ 506.686883] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 506.686883] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 506.687640] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 2c453256-9d51-46a3-b814-54223f5fc641, please check neutron logs for more information. [ 506.687640] env[61906]: Removing descriptor: 15 [ 506.687640] env[61906]: ERROR nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2c453256-9d51-46a3-b814-54223f5fc641, please check neutron logs for more information. [ 506.687640] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Traceback (most recent call last): [ 506.687640] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 506.687640] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] yield resources [ 506.687640] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 506.687640] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] self.driver.spawn(context, instance, image_meta, [ 506.687640] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 506.687640] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] self._vmops.spawn(context, instance, image_meta, injected_files, [ 506.687640] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 506.687640] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] vm_ref = self.build_virtual_machine(instance, [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] vif_infos = vmwarevif.get_vif_info(self._session, [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] for vif in network_info: [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] return self._sync_wrapper(fn, *args, **kwargs) [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] self.wait() [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] self[:] = self._gt.wait() [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] return self._exit_event.wait() [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 506.687985] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] result = hub.switch() [ 506.688408] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 506.688408] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] return self.greenlet.switch() [ 506.688408] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 506.688408] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] result = function(*args, **kwargs) [ 506.688408] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 506.688408] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] return func(*args, **kwargs) [ 506.688408] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 506.688408] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] raise e [ 506.688408] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 506.688408] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] nwinfo = self.network_api.allocate_for_instance( [ 506.688408] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 506.688408] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] created_port_ids = self._update_ports_for_instance( [ 506.688408] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 506.688898] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] with excutils.save_and_reraise_exception(): [ 506.688898] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 506.688898] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] self.force_reraise() [ 506.688898] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 506.688898] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] raise self.value [ 506.688898] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 506.688898] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] updated_port = self._update_port( [ 506.688898] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 506.688898] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] _ensure_no_port_binding_failure(port) [ 506.688898] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 506.688898] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] raise exception.PortBindingFailed(port_id=port['id']) [ 506.688898] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] nova.exception.PortBindingFailed: Binding failed for port 2c453256-9d51-46a3-b814-54223f5fc641, please check neutron logs for more information. [ 506.688898] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] [ 506.689275] env[61906]: INFO nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Terminating instance [ 506.695025] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Acquiring lock "refresh_cache-a3e97bd0-34e0-45f2-a10b-f554e10ca710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 506.695025] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Acquired lock "refresh_cache-a3e97bd0-34e0-45f2-a10b-f554e10ca710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 506.695025] env[61906]: DEBUG nova.network.neutron [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 506.706549] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb07540-d033-4eba-9c4f-94c2c54e29ab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.714970] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa2d71c-acbb-4106-892c-c67820e64e90 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.746696] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ad43a3-99ef-4675-a112-0aaea8e07eec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.754500] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2b028e-f4d7-4325-bbe5-cdfab96bdcd9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.769528] env[61906]: DEBUG nova.compute.provider_tree [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 506.836962] env[61906]: DEBUG nova.network.neutron [req-ea54033b-4fdf-4b00-b1e3-7785982baa0c req-4fc974fa-be5e-4d30-a12a-2d239132e319 service nova] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 506.867581] env[61906]: DEBUG nova.compute.manager [req-c591b543-360d-4a57-ae11-48dfc1bdaa05 req-8c00a3ee-019d-4b8c-9a56-7517d604a30f service nova] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Received event network-changed-2c453256-9d51-46a3-b814-54223f5fc641 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 506.867846] env[61906]: DEBUG nova.compute.manager [req-c591b543-360d-4a57-ae11-48dfc1bdaa05 req-8c00a3ee-019d-4b8c-9a56-7517d604a30f service nova] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Refreshing instance network info cache due to event network-changed-2c453256-9d51-46a3-b814-54223f5fc641. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 506.867954] env[61906]: DEBUG oslo_concurrency.lockutils [req-c591b543-360d-4a57-ae11-48dfc1bdaa05 req-8c00a3ee-019d-4b8c-9a56-7517d604a30f service nova] Acquiring lock "refresh_cache-a3e97bd0-34e0-45f2-a10b-f554e10ca710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 506.891881] env[61906]: DEBUG nova.network.neutron [-] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 506.916839] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356430, 'name': ReconfigVM_Task, 'duration_secs': 0.763601} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 506.917308] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45/85effb81-494c-4bda-a093-6dde523d5f45.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 506.917980] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2acbbd88-9eb6-4329-92c3-89d785f23af1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.925160] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 506.925160] env[61906]: value = "task-1356431" [ 506.925160] env[61906]: _type = "Task" [ 506.925160] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 506.934311] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356431, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 507.054305] env[61906]: DEBUG nova.network.neutron [req-ea54033b-4fdf-4b00-b1e3-7785982baa0c req-4fc974fa-be5e-4d30-a12a-2d239132e319 service nova] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 507.056979] env[61906]: DEBUG nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 507.266912] env[61906]: DEBUG nova.network.neutron [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 507.272239] env[61906]: DEBUG nova.scheduler.client.report [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 507.395267] env[61906]: INFO nova.compute.manager [-] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Took 1.06 seconds to deallocate network for instance. [ 507.398263] env[61906]: DEBUG nova.compute.claims [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 507.398331] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 507.436276] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356431, 'name': Rename_Task, 'duration_secs': 0.147621} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 507.436559] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 507.436795] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2a14b4b-cefa-4a5c-9364-24c191d2defb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.448231] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 507.448231] env[61906]: value = "task-1356432" [ 507.448231] env[61906]: _type = "Task" [ 507.448231] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 507.455144] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356432, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 507.499732] env[61906]: DEBUG nova.network.neutron [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 507.552076] env[61906]: DEBUG nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 507.561557] env[61906]: DEBUG oslo_concurrency.lockutils [req-ea54033b-4fdf-4b00-b1e3-7785982baa0c req-4fc974fa-be5e-4d30-a12a-2d239132e319 service nova] Releasing lock "refresh_cache-c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 507.586135] env[61906]: DEBUG nova.virt.hardware [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 507.586424] env[61906]: DEBUG nova.virt.hardware [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 507.586580] env[61906]: DEBUG nova.virt.hardware [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 507.587535] env[61906]: DEBUG nova.virt.hardware [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 507.587725] env[61906]: DEBUG nova.virt.hardware [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 507.587877] env[61906]: DEBUG nova.virt.hardware [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 507.588097] env[61906]: DEBUG nova.virt.hardware [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 507.588510] env[61906]: DEBUG nova.virt.hardware [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 507.588724] env[61906]: DEBUG nova.virt.hardware [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 507.588890] env[61906]: DEBUG nova.virt.hardware [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 507.589077] env[61906]: DEBUG nova.virt.hardware [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 507.590271] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839d0e9e-d556-4b76-a2b2-697973f17ab9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.594885] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 507.600489] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec98ba4-946e-4b96-b045-382b59bc02f9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.780561] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.273s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 507.782930] env[61906]: DEBUG nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 507.785660] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.068s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 507.791120] env[61906]: INFO nova.compute.claims [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 507.792383] env[61906]: DEBUG nova.network.neutron [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Successfully created port: af4e9b5d-e4bf-4c0b-a45c-0e963d17f749 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 507.966129] env[61906]: DEBUG oslo_vmware.api [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356432, 'name': PowerOnVM_Task, 'duration_secs': 0.441335} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 507.966510] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 507.967210] env[61906]: INFO nova.compute.manager [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Took 9.70 seconds to spawn the instance on the hypervisor. [ 507.967210] env[61906]: DEBUG nova.compute.manager [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 507.967734] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a31660b7-1133-4bd1-a211-2fe38451f4e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.006292] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Releasing lock "refresh_cache-a3e97bd0-34e0-45f2-a10b-f554e10ca710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 508.007803] env[61906]: DEBUG nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 508.007803] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 508.007803] env[61906]: DEBUG oslo_concurrency.lockutils [req-c591b543-360d-4a57-ae11-48dfc1bdaa05 req-8c00a3ee-019d-4b8c-9a56-7517d604a30f service nova] Acquired lock "refresh_cache-a3e97bd0-34e0-45f2-a10b-f554e10ca710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 508.007803] env[61906]: DEBUG nova.network.neutron [req-c591b543-360d-4a57-ae11-48dfc1bdaa05 req-8c00a3ee-019d-4b8c-9a56-7517d604a30f service nova] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Refreshing network info cache for port 2c453256-9d51-46a3-b814-54223f5fc641 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 508.010775] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91e98cfa-ffb6-48cc-8a6c-ad270dbaa838 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.021669] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff38cb32-643a-4088-b9b1-bdab15580641 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.047419] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a3e97bd0-34e0-45f2-a10b-f554e10ca710 could not be found. [ 508.047726] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 508.047969] env[61906]: INFO nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Took 0.04 seconds to destroy the instance on the hypervisor. [ 508.048235] env[61906]: DEBUG oslo.service.loopingcall [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 508.048462] env[61906]: DEBUG nova.compute.manager [-] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 508.048605] env[61906]: DEBUG nova.network.neutron [-] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 508.237118] env[61906]: DEBUG nova.network.neutron [-] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 508.298376] env[61906]: DEBUG nova.compute.utils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 508.308520] env[61906]: DEBUG nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 508.308520] env[61906]: DEBUG nova.network.neutron [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 508.490450] env[61906]: INFO nova.compute.manager [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Took 16.07 seconds to build instance. [ 508.497498] env[61906]: DEBUG nova.policy [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca80129d492e4d68b96ef14a63336ef3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb85e88fd4f54e0db2ff131f81137f64', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 508.630398] env[61906]: DEBUG nova.network.neutron [req-c591b543-360d-4a57-ae11-48dfc1bdaa05 req-8c00a3ee-019d-4b8c-9a56-7517d604a30f service nova] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 508.637342] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 508.637720] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 508.637813] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 508.637936] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 508.742200] env[61906]: DEBUG nova.network.neutron [-] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 508.809715] env[61906]: DEBUG nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 508.986292] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf04dab-c88a-4516-b04f-8a1a77b7fce2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.997711] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1968775e-afc4-4d68-97eb-5df00039230c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.002993] env[61906]: DEBUG oslo_concurrency.lockutils [None req-56054d9b-22f2-4b60-b0e1-a2e6fe2cf23a tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Lock "85effb81-494c-4bda-a093-6dde523d5f45" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.595s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 509.038857] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662c9cf7-0f6d-4c7d-ba0a-f6ee7f71e715 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.054565] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ca0c48-6fba-42f8-ad48-733727bbfdb8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.069021] env[61906]: DEBUG nova.compute.provider_tree [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 509.118660] env[61906]: DEBUG nova.network.neutron [req-c591b543-360d-4a57-ae11-48dfc1bdaa05 req-8c00a3ee-019d-4b8c-9a56-7517d604a30f service nova] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 509.142786] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 509.143106] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 509.144137] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 509.144137] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 509.144137] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 509.144399] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 509.144547] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 509.218529] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "refresh_cache-85effb81-494c-4bda-a093-6dde523d5f45" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 509.218677] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquired lock "refresh_cache-85effb81-494c-4bda-a093-6dde523d5f45" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 509.218825] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 509.219043] env[61906]: DEBUG nova.objects.instance [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lazy-loading 'info_cache' on Instance uuid 85effb81-494c-4bda-a093-6dde523d5f45 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 509.246849] env[61906]: INFO nova.compute.manager [-] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Took 1.20 seconds to deallocate network for instance. [ 509.247925] env[61906]: DEBUG nova.compute.claims [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 509.248270] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 509.365358] env[61906]: DEBUG nova.compute.manager [req-1b2f3f28-def4-484d-8f7e-ea94ba7223dc req-f81f84d8-e7ea-42ee-911d-6700d9c453ea service nova] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Received event network-vif-deleted-e27c9e5a-bf20-4fde-aa0d-0f7859533c1c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 509.586611] env[61906]: DEBUG nova.scheduler.client.report [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 509.621728] env[61906]: DEBUG oslo_concurrency.lockutils [req-c591b543-360d-4a57-ae11-48dfc1bdaa05 req-8c00a3ee-019d-4b8c-9a56-7517d604a30f service nova] Releasing lock "refresh_cache-a3e97bd0-34e0-45f2-a10b-f554e10ca710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 509.802586] env[61906]: DEBUG nova.network.neutron [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Successfully created port: 61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 509.820636] env[61906]: DEBUG nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 509.860708] env[61906]: DEBUG nova.virt.hardware [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 509.860708] env[61906]: DEBUG nova.virt.hardware [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 509.860708] env[61906]: DEBUG nova.virt.hardware [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 509.860864] env[61906]: DEBUG nova.virt.hardware [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 509.860864] env[61906]: DEBUG nova.virt.hardware [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 509.860952] env[61906]: DEBUG nova.virt.hardware [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 509.865942] env[61906]: DEBUG nova.virt.hardware [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 509.865942] env[61906]: DEBUG nova.virt.hardware [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 509.865942] env[61906]: DEBUG nova.virt.hardware [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 509.865942] env[61906]: DEBUG nova.virt.hardware [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 509.865942] env[61906]: DEBUG nova.virt.hardware [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 509.866195] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c93e4f9-0d18-4ec7-8fbb-44944e58edb8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.876864] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad2d9dc-8fae-492e-83db-f0414d6a4110 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.018791] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Acquiring lock "9f6864a2-3e54-41cb-b858-16e6e350a2af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.019087] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Lock "9f6864a2-3e54-41cb-b858-16e6e350a2af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.086228] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.300s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 510.086228] env[61906]: DEBUG nova.compute.manager [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 510.088753] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 7.441s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.306330] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 510.521371] env[61906]: DEBUG nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 510.596929] env[61906]: DEBUG nova.compute.utils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 510.606373] env[61906]: DEBUG nova.compute.manager [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 510.621088] env[61906]: DEBUG nova.compute.manager [req-98436378-b828-4118-ada9-c9d0ae5e2e59 req-bd981c90-590f-4e24-9960-1e67f2f7e353 service nova] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Received event network-vif-deleted-2c453256-9d51-46a3-b814-54223f5fc641 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 510.715902] env[61906]: ERROR nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5fbe7162-22fd-463d-a62c-1725806badc1, please check neutron logs for more information. [ 510.715902] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 510.715902] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 510.715902] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 510.715902] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 510.715902] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 510.715902] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 510.715902] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 510.715902] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 510.715902] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 510.715902] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 510.715902] env[61906]: ERROR nova.compute.manager raise self.value [ 510.715902] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 510.715902] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 510.715902] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 510.715902] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 510.717308] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 510.717308] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 510.717308] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5fbe7162-22fd-463d-a62c-1725806badc1, please check neutron logs for more information. [ 510.717308] env[61906]: ERROR nova.compute.manager [ 510.717308] env[61906]: Traceback (most recent call last): [ 510.717308] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 510.717308] env[61906]: listener.cb(fileno) [ 510.717308] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 510.717308] env[61906]: result = function(*args, **kwargs) [ 510.717308] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 510.717308] env[61906]: return func(*args, **kwargs) [ 510.717308] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 510.717308] env[61906]: raise e [ 510.717308] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 510.717308] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 510.717308] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 510.717308] env[61906]: created_port_ids = self._update_ports_for_instance( [ 510.717308] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 510.717308] env[61906]: with excutils.save_and_reraise_exception(): [ 510.717308] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 510.717308] env[61906]: self.force_reraise() [ 510.717308] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 510.717308] env[61906]: raise self.value [ 510.717308] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 510.717308] env[61906]: updated_port = self._update_port( [ 510.717308] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 510.717308] env[61906]: _ensure_no_port_binding_failure(port) [ 510.717308] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 510.717308] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 510.718371] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 5fbe7162-22fd-463d-a62c-1725806badc1, please check neutron logs for more information. [ 510.718371] env[61906]: Removing descriptor: 16 [ 510.718452] env[61906]: ERROR nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5fbe7162-22fd-463d-a62c-1725806badc1, please check neutron logs for more information. [ 510.718452] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Traceback (most recent call last): [ 510.718452] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 510.718452] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] yield resources [ 510.718452] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 510.718452] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] self.driver.spawn(context, instance, image_meta, [ 510.718452] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 510.718452] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 510.718452] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 510.718452] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] vm_ref = self.build_virtual_machine(instance, [ 510.718452] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 510.719771] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] vif_infos = vmwarevif.get_vif_info(self._session, [ 510.719771] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 510.719771] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] for vif in network_info: [ 510.719771] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 510.719771] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] return self._sync_wrapper(fn, *args, **kwargs) [ 510.719771] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 510.719771] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] self.wait() [ 510.719771] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 510.719771] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] self[:] = self._gt.wait() [ 510.719771] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 510.719771] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] return self._exit_event.wait() [ 510.719771] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 510.719771] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] result = hub.switch() [ 510.720168] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 510.720168] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] return self.greenlet.switch() [ 510.720168] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 510.720168] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] result = function(*args, **kwargs) [ 510.720168] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 510.720168] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] return func(*args, **kwargs) [ 510.720168] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 510.720168] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] raise e [ 510.720168] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 510.720168] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] nwinfo = self.network_api.allocate_for_instance( [ 510.720168] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 510.720168] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] created_port_ids = self._update_ports_for_instance( [ 510.720168] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 510.720521] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] with excutils.save_and_reraise_exception(): [ 510.720521] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 510.720521] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] self.force_reraise() [ 510.720521] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 510.720521] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] raise self.value [ 510.720521] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 510.720521] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] updated_port = self._update_port( [ 510.720521] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 510.720521] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] _ensure_no_port_binding_failure(port) [ 510.720521] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 510.720521] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] raise exception.PortBindingFailed(port_id=port['id']) [ 510.720521] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] nova.exception.PortBindingFailed: Binding failed for port 5fbe7162-22fd-463d-a62c-1725806badc1, please check neutron logs for more information. [ 510.720521] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] [ 510.720913] env[61906]: INFO nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Terminating instance [ 510.727794] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquiring lock "refresh_cache-5f417757-4295-4faa-83d2-1ad2d9c04ad6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 510.727794] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquired lock "refresh_cache-5f417757-4295-4faa-83d2-1ad2d9c04ad6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 510.727794] env[61906]: DEBUG nova.network.neutron [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 510.851934] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0538ab47-db95-4214-b089-379bbfff7c05 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.860337] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb38a724-196f-430c-8152-646a20100782 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.893360] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d600c37-4600-4a0a-9efa-8f9bc1cd0c47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.901071] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1777b31-7eb6-4866-9e1c-108e1a2ce308 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.915247] env[61906]: DEBUG nova.compute.provider_tree [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 511.050497] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.105062] env[61906]: DEBUG nova.compute.manager [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 511.126752] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 511.279040] env[61906]: DEBUG nova.network.neutron [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 511.419628] env[61906]: DEBUG nova.scheduler.client.report [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 511.587517] env[61906]: DEBUG nova.network.neutron [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 511.631897] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Releasing lock "refresh_cache-85effb81-494c-4bda-a093-6dde523d5f45" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 511.632157] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 511.632390] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 511.632668] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 511.633219] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 511.633424] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 511.633611] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 511.633788] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 511.927108] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.838s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 511.927786] env[61906]: ERROR nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7f55ae0f-f69b-416e-9508-5c259ddb945d, please check neutron logs for more information. [ 511.927786] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Traceback (most recent call last): [ 511.927786] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 511.927786] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] self.driver.spawn(context, instance, image_meta, [ 511.927786] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 511.927786] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 511.927786] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 511.927786] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] vm_ref = self.build_virtual_machine(instance, [ 511.927786] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 511.927786] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] vif_infos = vmwarevif.get_vif_info(self._session, [ 511.927786] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 511.928243] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] for vif in network_info: [ 511.928243] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 511.928243] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] return self._sync_wrapper(fn, *args, **kwargs) [ 511.928243] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 511.928243] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] self.wait() [ 511.928243] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 511.928243] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] self[:] = self._gt.wait() [ 511.928243] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 511.928243] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] return self._exit_event.wait() [ 511.928243] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 511.928243] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] result = hub.switch() [ 511.928243] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 511.928243] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] return self.greenlet.switch() [ 511.928572] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 511.928572] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] result = function(*args, **kwargs) [ 511.928572] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 511.928572] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] return func(*args, **kwargs) [ 511.928572] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 511.928572] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] raise e [ 511.928572] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 511.928572] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] nwinfo = self.network_api.allocate_for_instance( [ 511.928572] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 511.928572] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] created_port_ids = self._update_ports_for_instance( [ 511.928572] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 511.928572] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] with excutils.save_and_reraise_exception(): [ 511.928572] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 511.928924] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] self.force_reraise() [ 511.928924] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 511.928924] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] raise self.value [ 511.928924] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 511.928924] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] updated_port = self._update_port( [ 511.928924] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 511.928924] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] _ensure_no_port_binding_failure(port) [ 511.928924] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 511.928924] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] raise exception.PortBindingFailed(port_id=port['id']) [ 511.928924] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] nova.exception.PortBindingFailed: Binding failed for port 7f55ae0f-f69b-416e-9508-5c259ddb945d, please check neutron logs for more information. [ 511.928924] env[61906]: ERROR nova.compute.manager [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] [ 511.929223] env[61906]: DEBUG nova.compute.utils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Binding failed for port 7f55ae0f-f69b-416e-9508-5c259ddb945d, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 511.935881] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 4.537s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.949791] env[61906]: DEBUG nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Build of instance 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3 was re-scheduled: Binding failed for port 7f55ae0f-f69b-416e-9508-5c259ddb945d, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 511.949791] env[61906]: DEBUG nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 511.949791] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Acquiring lock "refresh_cache-2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 511.949791] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Acquired lock "refresh_cache-2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 511.950112] env[61906]: DEBUG nova.network.neutron [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 511.996464] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Acquiring lock "4dc3af05-b670-4f7c-bb9f-d1fa2e4180db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.996622] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Lock "4dc3af05-b670-4f7c-bb9f-d1fa2e4180db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.090059] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Releasing lock "refresh_cache-5f417757-4295-4faa-83d2-1ad2d9c04ad6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 512.091145] env[61906]: DEBUG nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 512.091145] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 512.091145] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9730e4e3-c60d-4232-b8b8-d475a270c43c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.103235] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1be628-cdf3-40c1-a5fe-bd3fadcb869d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.115131] env[61906]: DEBUG nova.compute.manager [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 512.125709] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5f417757-4295-4faa-83d2-1ad2d9c04ad6 could not be found. [ 512.125950] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 512.126167] env[61906]: INFO nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 512.126442] env[61906]: DEBUG oslo.service.loopingcall [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 512.126716] env[61906]: DEBUG nova.compute.manager [-] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 512.126808] env[61906]: DEBUG nova.network.neutron [-] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 512.137873] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Getting list of instances from cluster (obj){ [ 512.137873] env[61906]: value = "domain-c8" [ 512.137873] env[61906]: _type = "ClusterComputeResource" [ 512.137873] env[61906]: } {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 512.138932] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80997eb9-1951-4b68-a8cc-6c2fae03944d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.153336] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Got total of 1 instances {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 512.153628] env[61906]: WARNING nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] While synchronizing instance power states, found 7 instances in the database and 1 instances on the hypervisor. [ 512.153628] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Triggering sync for uuid 85effb81-494c-4bda-a093-6dde523d5f45 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 512.154704] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Triggering sync for uuid c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 512.154704] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Triggering sync for uuid a3e97bd0-34e0-45f2-a10b-f554e10ca710 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 512.154704] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Triggering sync for uuid 5f417757-4295-4faa-83d2-1ad2d9c04ad6 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 512.154704] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Triggering sync for uuid 17b85f46-4660-4af6-825d-3f2e7ea931ca {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 512.154704] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Triggering sync for uuid 2f64e954-bf3e-464f-a72e-d83f0783af5c {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 512.154704] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Triggering sync for uuid 7f4bb9fb-fa64-451d-9876-7e4314c13a64 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 512.157167] env[61906]: DEBUG nova.virt.hardware [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 512.157167] env[61906]: DEBUG nova.virt.hardware [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 512.157347] env[61906]: DEBUG nova.virt.hardware [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 512.157407] env[61906]: DEBUG nova.virt.hardware [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 512.157983] env[61906]: DEBUG nova.virt.hardware [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 512.157983] env[61906]: DEBUG nova.virt.hardware [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 512.157983] env[61906]: DEBUG nova.virt.hardware [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 512.158134] env[61906]: DEBUG nova.virt.hardware [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 512.158591] env[61906]: DEBUG nova.virt.hardware [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 512.158591] env[61906]: DEBUG nova.virt.hardware [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 512.158591] env[61906]: DEBUG nova.virt.hardware [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 512.158814] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "85effb81-494c-4bda-a093-6dde523d5f45" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.158997] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "85effb81-494c-4bda-a093-6dde523d5f45" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.159268] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.159511] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "a3e97bd0-34e0-45f2-a10b-f554e10ca710" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.159758] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "5f417757-4295-4faa-83d2-1ad2d9c04ad6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.159839] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "17b85f46-4660-4af6-825d-3f2e7ea931ca" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.160072] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "2f64e954-bf3e-464f-a72e-d83f0783af5c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.164356] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "7f4bb9fb-fa64-451d-9876-7e4314c13a64" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.164567] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 512.164835] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 512.166081] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df3bbd0-e9f5-45ba-8a97-61b2003cf41f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.172674] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce2f260-2995-4bf8-9b60-565a225d5e47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.178816] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 512.181868] env[61906]: DEBUG nova.network.neutron [-] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 512.191808] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcfd4e4e-8d66-462f-8bbe-51ae4feb98d6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.211992] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 512.217890] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Creating folder: Project (f9669395a0794fad977bab1fb81ee0d2). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 512.218793] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35fc7354-96e1-4086-9c70-72c1d43bae41 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.228934] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Created folder: Project (f9669395a0794fad977bab1fb81ee0d2) in parent group-v288914. [ 512.229142] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Creating folder: Instances. Parent ref: group-v288918. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 512.229368] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e5cedd63-a21d-4ec7-8138-6136bd8e8750 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.238796] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Created folder: Instances in parent group-v288918. [ 512.239045] env[61906]: DEBUG oslo.service.loopingcall [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 512.239232] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 512.239440] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4f4f0ec2-3fc9-4c61-9d70-329307e43640 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.258825] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 512.258825] env[61906]: value = "task-1356435" [ 512.258825] env[61906]: _type = "Task" [ 512.258825] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.270169] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356435, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.465532] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Acquiring lock "e23fd777-66f7-4ee1-bdcb-fe19c55f955c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.465794] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Lock "e23fd777-66f7-4ee1-bdcb-fe19c55f955c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.499510] env[61906]: DEBUG nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 512.515690] env[61906]: DEBUG nova.network.neutron [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 512.639555] env[61906]: INFO nova.compute.manager [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Rebuilding instance [ 512.663529] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e082526-e0fb-4de9-b619-ba638e54f3ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.684862] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249c22d2-9720-4ae1-b791-2c77adcafd74 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.689492] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.689791] env[61906]: DEBUG nova.network.neutron [-] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 512.727867] env[61906]: INFO nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] During sync_power_state the instance has a pending task (rebuilding). Skip. [ 512.727951] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "85effb81-494c-4bda-a093-6dde523d5f45" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.569s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 512.736274] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77cf143d-9e32-43af-84de-e1dbce429338 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.748609] env[61906]: DEBUG nova.compute.manager [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 512.749829] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de43397-2a38-4a88-8d8b-bb42ec2a2406 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.760664] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07cb4a4-3bfd-459a-a851-56eef3b8f608 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.795113] env[61906]: DEBUG nova.compute.provider_tree [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 512.796934] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356435, 'name': CreateVM_Task, 'duration_secs': 0.467451} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.797820] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 512.797820] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 512.798380] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 512.798380] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 512.798482] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b271ee76-84f9-4f74-b660-8d08aceaedfb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.803737] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Waiting for the task: (returnval){ [ 512.803737] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522f464e-44b4-e19d-4026-a4cea6d8f13c" [ 512.803737] env[61906]: _type = "Task" [ 512.803737] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.818382] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522f464e-44b4-e19d-4026-a4cea6d8f13c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.922376] env[61906]: ERROR nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port af4e9b5d-e4bf-4c0b-a45c-0e963d17f749, please check neutron logs for more information. [ 512.922376] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 512.922376] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 512.922376] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 512.922376] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 512.922376] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 512.922376] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 512.922376] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 512.922376] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 512.922376] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 512.922376] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 512.922376] env[61906]: ERROR nova.compute.manager raise self.value [ 512.922376] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 512.922376] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 512.922376] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 512.922376] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 512.922828] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 512.922828] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 512.922828] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port af4e9b5d-e4bf-4c0b-a45c-0e963d17f749, please check neutron logs for more information. [ 512.922828] env[61906]: ERROR nova.compute.manager [ 512.922828] env[61906]: Traceback (most recent call last): [ 512.922828] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 512.922828] env[61906]: listener.cb(fileno) [ 512.922828] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 512.922828] env[61906]: result = function(*args, **kwargs) [ 512.922828] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 512.922828] env[61906]: return func(*args, **kwargs) [ 512.922828] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 512.922828] env[61906]: raise e [ 512.922828] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 512.922828] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 512.922828] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 512.922828] env[61906]: created_port_ids = self._update_ports_for_instance( [ 512.922828] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 512.922828] env[61906]: with excutils.save_and_reraise_exception(): [ 512.922828] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 512.922828] env[61906]: self.force_reraise() [ 512.922828] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 512.922828] env[61906]: raise self.value [ 512.922828] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 512.922828] env[61906]: updated_port = self._update_port( [ 512.922828] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 512.922828] env[61906]: _ensure_no_port_binding_failure(port) [ 512.922828] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 512.922828] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 512.923630] env[61906]: nova.exception.PortBindingFailed: Binding failed for port af4e9b5d-e4bf-4c0b-a45c-0e963d17f749, please check neutron logs for more information. [ 512.923630] env[61906]: Removing descriptor: 17 [ 512.923630] env[61906]: ERROR nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port af4e9b5d-e4bf-4c0b-a45c-0e963d17f749, please check neutron logs for more information. [ 512.923630] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Traceback (most recent call last): [ 512.923630] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 512.923630] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] yield resources [ 512.923630] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 512.923630] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] self.driver.spawn(context, instance, image_meta, [ 512.923630] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 512.923630] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] self._vmops.spawn(context, instance, image_meta, injected_files, [ 512.923630] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 512.923630] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] vm_ref = self.build_virtual_machine(instance, [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] vif_infos = vmwarevif.get_vif_info(self._session, [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] for vif in network_info: [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] return self._sync_wrapper(fn, *args, **kwargs) [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] self.wait() [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] self[:] = self._gt.wait() [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] return self._exit_event.wait() [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 512.923969] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] result = hub.switch() [ 512.924344] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 512.924344] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] return self.greenlet.switch() [ 512.924344] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 512.924344] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] result = function(*args, **kwargs) [ 512.924344] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 512.924344] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] return func(*args, **kwargs) [ 512.924344] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 512.924344] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] raise e [ 512.924344] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 512.924344] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] nwinfo = self.network_api.allocate_for_instance( [ 512.924344] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 512.924344] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] created_port_ids = self._update_ports_for_instance( [ 512.924344] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 512.924786] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] with excutils.save_and_reraise_exception(): [ 512.924786] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 512.924786] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] self.force_reraise() [ 512.924786] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 512.924786] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] raise self.value [ 512.924786] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 512.924786] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] updated_port = self._update_port( [ 512.924786] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 512.924786] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] _ensure_no_port_binding_failure(port) [ 512.924786] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 512.924786] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] raise exception.PortBindingFailed(port_id=port['id']) [ 512.924786] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] nova.exception.PortBindingFailed: Binding failed for port af4e9b5d-e4bf-4c0b-a45c-0e963d17f749, please check neutron logs for more information. [ 512.924786] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] [ 512.929170] env[61906]: INFO nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Terminating instance [ 512.930555] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Acquiring lock "refresh_cache-17b85f46-4660-4af6-825d-3f2e7ea931ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 512.930723] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Acquired lock "refresh_cache-17b85f46-4660-4af6-825d-3f2e7ea931ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 512.931013] env[61906]: DEBUG nova.network.neutron [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 512.992604] env[61906]: DEBUG nova.network.neutron [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 513.032325] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.081297] env[61906]: DEBUG nova.compute.manager [req-98b26110-66e7-4ad4-b862-3798a0ff0647 req-1d1abe2f-68d0-4850-9b59-5818b4a6a2b2 service nova] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Received event network-changed-5fbe7162-22fd-463d-a62c-1725806badc1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 513.081297] env[61906]: DEBUG nova.compute.manager [req-98b26110-66e7-4ad4-b862-3798a0ff0647 req-1d1abe2f-68d0-4850-9b59-5818b4a6a2b2 service nova] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Refreshing instance network info cache due to event network-changed-5fbe7162-22fd-463d-a62c-1725806badc1. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 513.082163] env[61906]: DEBUG oslo_concurrency.lockutils [req-98b26110-66e7-4ad4-b862-3798a0ff0647 req-1d1abe2f-68d0-4850-9b59-5818b4a6a2b2 service nova] Acquiring lock "refresh_cache-5f417757-4295-4faa-83d2-1ad2d9c04ad6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 513.082577] env[61906]: DEBUG oslo_concurrency.lockutils [req-98b26110-66e7-4ad4-b862-3798a0ff0647 req-1d1abe2f-68d0-4850-9b59-5818b4a6a2b2 service nova] Acquired lock "refresh_cache-5f417757-4295-4faa-83d2-1ad2d9c04ad6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 513.082931] env[61906]: DEBUG nova.network.neutron [req-98b26110-66e7-4ad4-b862-3798a0ff0647 req-1d1abe2f-68d0-4850-9b59-5818b4a6a2b2 service nova] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Refreshing network info cache for port 5fbe7162-22fd-463d-a62c-1725806badc1 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 513.192477] env[61906]: INFO nova.compute.manager [-] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Took 1.07 seconds to deallocate network for instance. [ 513.196404] env[61906]: DEBUG nova.compute.claims [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 513.196827] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.280507] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 513.282300] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-560dfe21-8813-40a6-8c70-6e1a79d4b87c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.290463] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 513.290463] env[61906]: value = "task-1356436" [ 513.290463] env[61906]: _type = "Task" [ 513.290463] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.298308] env[61906]: DEBUG nova.scheduler.client.report [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 513.307870] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356436, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.319106] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522f464e-44b4-e19d-4026-a4cea6d8f13c, 'name': SearchDatastore_Task, 'duration_secs': 0.011036} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.319650] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 513.319650] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 513.320031] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 513.320140] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 513.320329] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 513.320596] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-983b2657-bcc1-4fde-8e01-27d6df04eaf4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.330287] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 513.330483] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 513.331223] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4867cefa-6834-403c-a844-f901820b8594 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.341729] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Waiting for the task: (returnval){ [ 513.341729] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528c7c01-6c6a-fa51-5560-5c367cac5b80" [ 513.341729] env[61906]: _type = "Task" [ 513.341729] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.351534] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528c7c01-6c6a-fa51-5560-5c367cac5b80, 'name': SearchDatastore_Task, 'duration_secs': 0.008829} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.352352] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b551ea42-8165-4e05-a818-8b962e04cff6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.359568] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Waiting for the task: (returnval){ [ 513.359568] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b197cb-5246-d91a-b9b5-4f9cbefd4ddb" [ 513.359568] env[61906]: _type = "Task" [ 513.359568] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.373316] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b197cb-5246-d91a-b9b5-4f9cbefd4ddb, 'name': SearchDatastore_Task, 'duration_secs': 0.008486} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.373316] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 513.373316] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 7f4bb9fb-fa64-451d-9876-7e4314c13a64/7f4bb9fb-fa64-451d-9876-7e4314c13a64.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 513.373316] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb1a4dd0-333e-4e8c-a7bc-843e67f3b3dc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.379650] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Waiting for the task: (returnval){ [ 513.379650] env[61906]: value = "task-1356437" [ 513.379650] env[61906]: _type = "Task" [ 513.379650] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.388613] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356437, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.480539] env[61906]: DEBUG nova.network.neutron [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 513.500404] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Releasing lock "refresh_cache-2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 513.500404] env[61906]: DEBUG nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 513.500404] env[61906]: DEBUG nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 513.500404] env[61906]: DEBUG nova.network.neutron [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 513.550490] env[61906]: DEBUG nova.network.neutron [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 513.635029] env[61906]: DEBUG nova.network.neutron [req-98b26110-66e7-4ad4-b862-3798a0ff0647 req-1d1abe2f-68d0-4850-9b59-5818b4a6a2b2 service nova] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 513.710228] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Acquiring lock "6c354d85-0252-4eb9-81e2-550a403bdbc5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.710228] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Lock "6c354d85-0252-4eb9-81e2-550a403bdbc5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.776313] env[61906]: ERROR nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a, please check neutron logs for more information. [ 513.776313] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 513.776313] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 513.776313] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 513.776313] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 513.776313] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 513.776313] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 513.776313] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 513.776313] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 513.776313] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 513.776313] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 513.776313] env[61906]: ERROR nova.compute.manager raise self.value [ 513.776313] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 513.776313] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 513.776313] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 513.776313] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 513.777000] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 513.777000] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 513.777000] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a, please check neutron logs for more information. [ 513.777000] env[61906]: ERROR nova.compute.manager [ 513.777000] env[61906]: Traceback (most recent call last): [ 513.777000] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 513.777000] env[61906]: listener.cb(fileno) [ 513.777000] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 513.777000] env[61906]: result = function(*args, **kwargs) [ 513.777000] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 513.777000] env[61906]: return func(*args, **kwargs) [ 513.777000] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 513.777000] env[61906]: raise e [ 513.777000] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 513.777000] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 513.777000] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 513.777000] env[61906]: created_port_ids = self._update_ports_for_instance( [ 513.777000] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 513.777000] env[61906]: with excutils.save_and_reraise_exception(): [ 513.777000] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 513.777000] env[61906]: self.force_reraise() [ 513.777000] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 513.777000] env[61906]: raise self.value [ 513.777000] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 513.777000] env[61906]: updated_port = self._update_port( [ 513.777000] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 513.777000] env[61906]: _ensure_no_port_binding_failure(port) [ 513.777000] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 513.777000] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 513.778474] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a, please check neutron logs for more information. [ 513.778474] env[61906]: Removing descriptor: 15 [ 513.778474] env[61906]: ERROR nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a, please check neutron logs for more information. [ 513.778474] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Traceback (most recent call last): [ 513.778474] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 513.778474] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] yield resources [ 513.778474] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 513.778474] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] self.driver.spawn(context, instance, image_meta, [ 513.778474] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 513.778474] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 513.778474] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 513.778474] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] vm_ref = self.build_virtual_machine(instance, [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] vif_infos = vmwarevif.get_vif_info(self._session, [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] for vif in network_info: [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] return self._sync_wrapper(fn, *args, **kwargs) [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] self.wait() [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] self[:] = self._gt.wait() [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] return self._exit_event.wait() [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 513.779054] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] result = hub.switch() [ 513.779448] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 513.779448] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] return self.greenlet.switch() [ 513.779448] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 513.779448] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] result = function(*args, **kwargs) [ 513.779448] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 513.779448] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] return func(*args, **kwargs) [ 513.779448] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 513.779448] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] raise e [ 513.779448] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 513.779448] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] nwinfo = self.network_api.allocate_for_instance( [ 513.779448] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 513.779448] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] created_port_ids = self._update_ports_for_instance( [ 513.779448] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 513.779787] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] with excutils.save_and_reraise_exception(): [ 513.779787] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 513.779787] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] self.force_reraise() [ 513.779787] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 513.779787] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] raise self.value [ 513.779787] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 513.779787] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] updated_port = self._update_port( [ 513.779787] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 513.779787] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] _ensure_no_port_binding_failure(port) [ 513.779787] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 513.779787] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] raise exception.PortBindingFailed(port_id=port['id']) [ 513.779787] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] nova.exception.PortBindingFailed: Binding failed for port 61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a, please check neutron logs for more information. [ 513.779787] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] [ 513.780378] env[61906]: INFO nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Terminating instance [ 513.782345] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "refresh_cache-2f64e954-bf3e-464f-a72e-d83f0783af5c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 513.782509] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "refresh_cache-2f64e954-bf3e-464f-a72e-d83f0783af5c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 513.782567] env[61906]: DEBUG nova.network.neutron [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 513.811358] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.874s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 513.811358] env[61906]: ERROR nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e27c9e5a-bf20-4fde-aa0d-0f7859533c1c, please check neutron logs for more information. [ 513.811358] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Traceback (most recent call last): [ 513.811358] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 513.811358] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] self.driver.spawn(context, instance, image_meta, [ 513.811358] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 513.811358] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 513.811358] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 513.811358] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] vm_ref = self.build_virtual_machine(instance, [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] for vif in network_info: [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] return self._sync_wrapper(fn, *args, **kwargs) [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] self.wait() [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] self[:] = self._gt.wait() [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] return self._exit_event.wait() [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 513.811664] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] result = hub.switch() [ 513.812084] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 513.812084] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] return self.greenlet.switch() [ 513.812084] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 513.812084] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] result = function(*args, **kwargs) [ 513.812084] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 513.812084] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] return func(*args, **kwargs) [ 513.812084] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 513.812084] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] raise e [ 513.812084] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 513.812084] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] nwinfo = self.network_api.allocate_for_instance( [ 513.812084] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 513.812084] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] created_port_ids = self._update_ports_for_instance( [ 513.812084] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 513.812570] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] with excutils.save_and_reraise_exception(): [ 513.812570] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 513.812570] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] self.force_reraise() [ 513.812570] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 513.812570] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] raise self.value [ 513.812570] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 513.812570] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] updated_port = self._update_port( [ 513.812570] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 513.812570] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] _ensure_no_port_binding_failure(port) [ 513.812570] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 513.812570] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] raise exception.PortBindingFailed(port_id=port['id']) [ 513.812570] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] nova.exception.PortBindingFailed: Binding failed for port e27c9e5a-bf20-4fde-aa0d-0f7859533c1c, please check neutron logs for more information. [ 513.812570] env[61906]: ERROR nova.compute.manager [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] [ 513.812949] env[61906]: DEBUG nova.compute.utils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Binding failed for port e27c9e5a-bf20-4fde-aa0d-0f7859533c1c, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 513.812949] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356436, 'name': PowerOffVM_Task, 'duration_secs': 0.121918} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.812949] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.217s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.818916] env[61906]: INFO nova.compute.claims [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 513.820700] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 513.820700] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 513.821311] env[61906]: DEBUG nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Build of instance c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1 was re-scheduled: Binding failed for port e27c9e5a-bf20-4fde-aa0d-0f7859533c1c, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 513.821870] env[61906]: DEBUG nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 513.822199] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquiring lock "refresh_cache-c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 513.822361] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquired lock "refresh_cache-c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 513.822574] env[61906]: DEBUG nova.network.neutron [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 513.824934] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c5c0bf-c199-4ce3-b3bb-3cd317f6a238 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.839700] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 513.841604] env[61906]: DEBUG nova.network.neutron [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 513.843197] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-99270a67-0227-4d3b-81ac-e745beeeeddd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.859815] env[61906]: DEBUG nova.network.neutron [req-98b26110-66e7-4ad4-b862-3798a0ff0647 req-1d1abe2f-68d0-4850-9b59-5818b4a6a2b2 service nova] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 513.871334] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 513.871626] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 513.871768] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Deleting the datastore file [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 513.873108] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e5d69df2-1d8e-4bff-8c07-cee2e68da23b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.884721] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 513.884721] env[61906]: value = "task-1356439" [ 513.884721] env[61906]: _type = "Task" [ 513.884721] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.897944] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356439, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.908185] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356437, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503105} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.908455] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 7f4bb9fb-fa64-451d-9876-7e4314c13a64/7f4bb9fb-fa64-451d-9876-7e4314c13a64.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 513.908664] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 513.909239] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-66a9b2bf-1336-43b6-a314-5e9bd1722df6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.917394] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Waiting for the task: (returnval){ [ 513.917394] env[61906]: value = "task-1356440" [ 513.917394] env[61906]: _type = "Task" [ 513.917394] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.928338] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356440, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.054584] env[61906]: DEBUG nova.network.neutron [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 514.320828] env[61906]: DEBUG nova.network.neutron [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 514.341684] env[61906]: DEBUG nova.compute.manager [req-2ba14f98-a56e-4ba8-96e1-d889e6bac594 req-6c25ed02-3436-45bc-9ce4-c54c474dc624 service nova] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Received event network-changed-61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 514.343286] env[61906]: DEBUG nova.compute.manager [req-2ba14f98-a56e-4ba8-96e1-d889e6bac594 req-6c25ed02-3436-45bc-9ce4-c54c474dc624 service nova] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Refreshing instance network info cache due to event network-changed-61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 514.343647] env[61906]: DEBUG oslo_concurrency.lockutils [req-2ba14f98-a56e-4ba8-96e1-d889e6bac594 req-6c25ed02-3436-45bc-9ce4-c54c474dc624 service nova] Acquiring lock "refresh_cache-2f64e954-bf3e-464f-a72e-d83f0783af5c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 514.346918] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Releasing lock "refresh_cache-17b85f46-4660-4af6-825d-3f2e7ea931ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 514.347550] env[61906]: DEBUG nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 514.348057] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 514.348647] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e0a05753-2319-455c-b113-7781f10a4efc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.361753] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b80a642-da15-4369-b21b-7f8e3bd4c812 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.377731] env[61906]: DEBUG nova.network.neutron [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 514.379957] env[61906]: DEBUG oslo_concurrency.lockutils [req-98b26110-66e7-4ad4-b862-3798a0ff0647 req-1d1abe2f-68d0-4850-9b59-5818b4a6a2b2 service nova] Releasing lock "refresh_cache-5f417757-4295-4faa-83d2-1ad2d9c04ad6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 514.380237] env[61906]: DEBUG nova.compute.manager [req-98b26110-66e7-4ad4-b862-3798a0ff0647 req-1d1abe2f-68d0-4850-9b59-5818b4a6a2b2 service nova] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Received event network-vif-deleted-5fbe7162-22fd-463d-a62c-1725806badc1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 514.396021] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 17b85f46-4660-4af6-825d-3f2e7ea931ca could not be found. [ 514.396021] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 514.396021] env[61906]: INFO nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Took 0.05 seconds to destroy the instance on the hypervisor. [ 514.396021] env[61906]: DEBUG oslo.service.loopingcall [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 514.397788] env[61906]: DEBUG nova.compute.manager [-] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 514.397889] env[61906]: DEBUG nova.network.neutron [-] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 514.409727] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356439, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.103857} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 514.409727] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 514.409727] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 514.409994] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 514.430103] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356440, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094649} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 514.430103] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 514.430741] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa4e9d8-6dc2-45ea-bb00-ee1244df69fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.454165] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Reconfiguring VM instance instance-00000008 to attach disk [datastore2] 7f4bb9fb-fa64-451d-9876-7e4314c13a64/7f4bb9fb-fa64-451d-9876-7e4314c13a64.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 514.454481] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7accdbe5-74d5-47a3-b5c3-3bd690d6d7e9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.468990] env[61906]: DEBUG nova.network.neutron [-] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 514.476546] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Waiting for the task: (returnval){ [ 514.476546] env[61906]: value = "task-1356441" [ 514.476546] env[61906]: _type = "Task" [ 514.476546] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 514.482814] env[61906]: DEBUG nova.network.neutron [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 514.487916] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356441, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.561602] env[61906]: INFO nova.compute.manager [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] [instance: 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3] Took 1.06 seconds to deallocate network for instance. [ 514.624990] env[61906]: DEBUG nova.network.neutron [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 514.728378] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquiring lock "f708ec4a-7f1e-4bfa-b561-4b9bc5556d27" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.728779] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Lock "f708ec4a-7f1e-4bfa-b561-4b9bc5556d27" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.974478] env[61906]: DEBUG nova.network.neutron [-] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 514.989481] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356441, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.990685] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Releasing lock "refresh_cache-c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 514.991111] env[61906]: DEBUG nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 514.991477] env[61906]: DEBUG nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 514.991684] env[61906]: DEBUG nova.network.neutron [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 515.028267] env[61906]: DEBUG nova.network.neutron [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 515.129281] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "refresh_cache-2f64e954-bf3e-464f-a72e-d83f0783af5c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 515.129281] env[61906]: DEBUG nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 515.129281] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 515.130789] env[61906]: DEBUG oslo_concurrency.lockutils [req-2ba14f98-a56e-4ba8-96e1-d889e6bac594 req-6c25ed02-3436-45bc-9ce4-c54c474dc624 service nova] Acquired lock "refresh_cache-2f64e954-bf3e-464f-a72e-d83f0783af5c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 515.130789] env[61906]: DEBUG nova.network.neutron [req-2ba14f98-a56e-4ba8-96e1-d889e6bac594 req-6c25ed02-3436-45bc-9ce4-c54c474dc624 service nova] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Refreshing network info cache for port 61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 515.131532] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b13b7af1-5804-4686-a9cb-dae3d95e5b87 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.147313] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b775994a-4f1b-4407-81d0-43b892c6a207 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.179653] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2f64e954-bf3e-464f-a72e-d83f0783af5c could not be found. [ 515.179653] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 515.179847] env[61906]: INFO nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 515.180186] env[61906]: DEBUG oslo.service.loopingcall [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 515.180437] env[61906]: DEBUG nova.compute.manager [-] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 515.180543] env[61906]: DEBUG nova.network.neutron [-] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 515.184767] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720f9144-2d28-4033-afb1-445c858dd1c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.194035] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51455612-e56a-4913-b50f-fa7cb9468562 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.230630] env[61906]: DEBUG nova.network.neutron [-] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 515.232597] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22da097-b086-4e2c-a2fd-3ff757f6289b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.241189] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c86f3ba3-f6ca-474e-bd5c-64f0e0630522 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.258062] env[61906]: DEBUG nova.compute.provider_tree [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 515.478393] env[61906]: DEBUG nova.virt.hardware [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 515.478735] env[61906]: DEBUG nova.virt.hardware [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 515.478807] env[61906]: DEBUG nova.virt.hardware [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 515.479557] env[61906]: DEBUG nova.virt.hardware [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 515.481235] env[61906]: DEBUG nova.virt.hardware [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 515.481426] env[61906]: DEBUG nova.virt.hardware [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 515.481646] env[61906]: DEBUG nova.virt.hardware [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 515.481804] env[61906]: DEBUG nova.virt.hardware [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 515.481967] env[61906]: DEBUG nova.virt.hardware [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 515.482147] env[61906]: DEBUG nova.virt.hardware [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 515.482434] env[61906]: DEBUG nova.virt.hardware [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 515.483559] env[61906]: INFO nova.compute.manager [-] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Took 1.08 seconds to deallocate network for instance. [ 515.483817] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de50325-ae36-4904-a491-f53caf80cc5b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.497430] env[61906]: DEBUG nova.compute.claims [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 515.497636] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.509933] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90335929-6229-416b-bc15-137125752824 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.513078] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356441, 'name': ReconfigVM_Task, 'duration_secs': 0.710872} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 515.513340] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Reconfigured VM instance instance-00000008 to attach disk [datastore2] 7f4bb9fb-fa64-451d-9876-7e4314c13a64/7f4bb9fb-fa64-451d-9876-7e4314c13a64.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 515.514488] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-355af83e-0795-4910-9f1d-a0b161437adc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.525202] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 515.531156] env[61906]: DEBUG oslo.service.loopingcall [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 515.533206] env[61906]: DEBUG nova.network.neutron [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 515.534244] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 515.538169] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Waiting for the task: (returnval){ [ 515.538169] env[61906]: value = "task-1356442" [ 515.538169] env[61906]: _type = "Task" [ 515.538169] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.538578] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da4621a7-05fa-43f8-be8e-0230a660b29c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.567567] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356442, 'name': Rename_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.568938] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 515.568938] env[61906]: value = "task-1356443" [ 515.568938] env[61906]: _type = "Task" [ 515.568938] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.581238] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356443, 'name': CreateVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.610666] env[61906]: INFO nova.scheduler.client.report [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Deleted allocations for instance 2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3 [ 515.737636] env[61906]: DEBUG nova.network.neutron [-] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 515.762011] env[61906]: DEBUG nova.scheduler.client.report [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 515.826907] env[61906]: DEBUG nova.compute.manager [req-065128f2-1d39-4ad6-96d6-91fb76cf3997 req-50f6884d-8a73-42f1-ab12-34c5bc644416 service nova] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Received event network-changed-af4e9b5d-e4bf-4c0b-a45c-0e963d17f749 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 515.827118] env[61906]: DEBUG nova.compute.manager [req-065128f2-1d39-4ad6-96d6-91fb76cf3997 req-50f6884d-8a73-42f1-ab12-34c5bc644416 service nova] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Refreshing instance network info cache due to event network-changed-af4e9b5d-e4bf-4c0b-a45c-0e963d17f749. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 515.827335] env[61906]: DEBUG oslo_concurrency.lockutils [req-065128f2-1d39-4ad6-96d6-91fb76cf3997 req-50f6884d-8a73-42f1-ab12-34c5bc644416 service nova] Acquiring lock "refresh_cache-17b85f46-4660-4af6-825d-3f2e7ea931ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 515.827471] env[61906]: DEBUG oslo_concurrency.lockutils [req-065128f2-1d39-4ad6-96d6-91fb76cf3997 req-50f6884d-8a73-42f1-ab12-34c5bc644416 service nova] Acquired lock "refresh_cache-17b85f46-4660-4af6-825d-3f2e7ea931ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 515.827675] env[61906]: DEBUG nova.network.neutron [req-065128f2-1d39-4ad6-96d6-91fb76cf3997 req-50f6884d-8a73-42f1-ab12-34c5bc644416 service nova] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Refreshing network info cache for port af4e9b5d-e4bf-4c0b-a45c-0e963d17f749 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 515.874170] env[61906]: DEBUG nova.network.neutron [req-2ba14f98-a56e-4ba8-96e1-d889e6bac594 req-6c25ed02-3436-45bc-9ce4-c54c474dc624 service nova] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 516.011008] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquiring lock "ec81c4ec-07c6-4b1f-8555-69902cff4765" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.011264] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Lock "ec81c4ec-07c6-4b1f-8555-69902cff4765" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.040215] env[61906]: INFO nova.compute.manager [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] Took 1.05 seconds to deallocate network for instance. [ 516.073746] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356442, 'name': Rename_Task, 'duration_secs': 0.157131} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 516.079155] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 516.080566] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-84495638-5686-4b00-a55e-74d3baa7c7db {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.089859] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356443, 'name': CreateVM_Task, 'duration_secs': 0.328388} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 516.091119] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 516.091649] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Waiting for the task: (returnval){ [ 516.091649] env[61906]: value = "task-1356444" [ 516.091649] env[61906]: _type = "Task" [ 516.091649] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.092591] env[61906]: DEBUG oslo_concurrency.lockutils [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 516.092591] env[61906]: DEBUG oslo_concurrency.lockutils [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 516.092591] env[61906]: DEBUG oslo_concurrency.lockutils [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 516.092752] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f88159b8-e582-4fc0-b504-b98fc2405b4f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.106167] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356444, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.107855] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 516.107855] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]526e6be8-7959-6518-19d1-23704c26853f" [ 516.107855] env[61906]: _type = "Task" [ 516.107855] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.124126] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a009a21b-093e-4aca-90e0-424853bb5513 tempest-ImagesNegativeTestJSON-681998903 tempest-ImagesNegativeTestJSON-681998903-project-member] Lock "2e14f0cb-bf5e-4aa2-8b18-a4905705a3b3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.437s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 516.124126] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]526e6be8-7959-6518-19d1-23704c26853f, 'name': SearchDatastore_Task, 'duration_secs': 0.010093} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 516.124126] env[61906]: DEBUG oslo_concurrency.lockutils [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 516.124126] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 516.124586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 516.124586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 516.124586] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 516.125160] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1749e838-d3f7-44d9-a455-9009407d0ae2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.133788] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 516.134070] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 516.134961] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4eb217df-879e-4340-8215-042bc032c3a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.141953] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 516.141953] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52e38c35-cb88-7280-227e-2812e6fcbb79" [ 516.141953] env[61906]: _type = "Task" [ 516.141953] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.151738] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52e38c35-cb88-7280-227e-2812e6fcbb79, 'name': SearchDatastore_Task, 'duration_secs': 0.007644} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 516.155970] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3d8afd0-3666-4231-a735-59ffcad9c02c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.162268] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 516.162268] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528e53d4-8fcc-e7f8-3e5f-3829f9bd986d" [ 516.162268] env[61906]: _type = "Task" [ 516.162268] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.175628] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528e53d4-8fcc-e7f8-3e5f-3829f9bd986d, 'name': SearchDatastore_Task, 'duration_secs': 0.008413} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 516.178978] env[61906]: DEBUG oslo_concurrency.lockutils [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 516.178978] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45/85effb81-494c-4bda-a093-6dde523d5f45.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 516.178978] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aa335158-76b4-4717-af5f-c6b1f8b75e1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.182589] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 516.182589] env[61906]: value = "task-1356445" [ 516.182589] env[61906]: _type = "Task" [ 516.182589] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.190538] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356445, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.197782] env[61906]: DEBUG nova.network.neutron [req-2ba14f98-a56e-4ba8-96e1-d889e6bac594 req-6c25ed02-3436-45bc-9ce4-c54c474dc624 service nova] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 516.243031] env[61906]: INFO nova.compute.manager [-] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Took 1.06 seconds to deallocate network for instance. [ 516.245538] env[61906]: DEBUG nova.compute.claims [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 516.245538] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.267893] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 516.272047] env[61906]: DEBUG nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 516.273143] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 7.025s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.370936] env[61906]: DEBUG nova.network.neutron [req-065128f2-1d39-4ad6-96d6-91fb76cf3997 req-50f6884d-8a73-42f1-ab12-34c5bc644416 service nova] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 516.611876] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356444, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.629742] env[61906]: DEBUG nova.compute.manager [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.684497] env[61906]: DEBUG nova.network.neutron [req-065128f2-1d39-4ad6-96d6-91fb76cf3997 req-50f6884d-8a73-42f1-ab12-34c5bc644416 service nova] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 516.697839] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356445, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498337} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 516.697962] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45/85effb81-494c-4bda-a093-6dde523d5f45.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 516.699032] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 516.699331] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea50bc60-7ce3-404b-9d12-c3c3efdb8664 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.701932] env[61906]: DEBUG oslo_concurrency.lockutils [req-2ba14f98-a56e-4ba8-96e1-d889e6bac594 req-6c25ed02-3436-45bc-9ce4-c54c474dc624 service nova] Releasing lock "refresh_cache-2f64e954-bf3e-464f-a72e-d83f0783af5c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 516.709876] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 516.709876] env[61906]: value = "task-1356446" [ 516.709876] env[61906]: _type = "Task" [ 516.709876] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.721354] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356446, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.783545] env[61906]: DEBUG nova.compute.utils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 516.787951] env[61906]: DEBUG nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 516.788066] env[61906]: DEBUG nova.network.neutron [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 516.954701] env[61906]: DEBUG nova.policy [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7fe3db47470840beafe3d5bcf1bb9f74', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '536493702a8849f89aa5f8e7482cec9b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 517.028283] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0719dd-78e4-4a53-b645-e34f3f9d119c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.036872] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf616264-2c3b-49cf-baae-3be1ea46f88c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.076276] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20893dd-7dcf-4b70-9200-4cd75ee85584 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.084127] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04218923-c9cd-4d54-96d4-bb6c4b5c110b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.101233] env[61906]: DEBUG nova.compute.provider_tree [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 517.114124] env[61906]: DEBUG oslo_vmware.api [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356444, 'name': PowerOnVM_Task, 'duration_secs': 0.973029} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 517.114230] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 517.115861] env[61906]: INFO nova.compute.manager [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Took 5.00 seconds to spawn the instance on the hypervisor. [ 517.115861] env[61906]: DEBUG nova.compute.manager [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 517.115861] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc38782-09f7-4d0f-9522-fac313b2daec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.141342] env[61906]: INFO nova.scheduler.client.report [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Deleted allocations for instance c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1 [ 517.162469] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.192300] env[61906]: DEBUG oslo_concurrency.lockutils [req-065128f2-1d39-4ad6-96d6-91fb76cf3997 req-50f6884d-8a73-42f1-ab12-34c5bc644416 service nova] Releasing lock "refresh_cache-17b85f46-4660-4af6-825d-3f2e7ea931ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 517.192595] env[61906]: DEBUG nova.compute.manager [req-065128f2-1d39-4ad6-96d6-91fb76cf3997 req-50f6884d-8a73-42f1-ab12-34c5bc644416 service nova] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Received event network-vif-deleted-af4e9b5d-e4bf-4c0b-a45c-0e963d17f749 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 517.222316] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356446, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075083} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 517.222656] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 517.223530] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e54a0ad8-02c2-426a-a08b-264df584c4c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.249291] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45/85effb81-494c-4bda-a093-6dde523d5f45.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 517.249665] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f1d8d0b-cb4b-4af7-8b83-cabc08e60201 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.271761] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 517.271761] env[61906]: value = "task-1356447" [ 517.271761] env[61906]: _type = "Task" [ 517.271761] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 517.281778] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356447, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.288826] env[61906]: DEBUG nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 517.295122] env[61906]: DEBUG nova.compute.manager [req-65a7ec10-27c5-46b1-b0eb-ead267690f61 req-8406a738-9ff1-45b2-aa58-77ff16d9e8ca service nova] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Received event network-vif-deleted-61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 517.609407] env[61906]: DEBUG nova.scheduler.client.report [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 517.637655] env[61906]: INFO nova.compute.manager [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Took 16.95 seconds to build instance. [ 517.649937] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a130e620-7dea-41cb-91d3-0ff4155a0ce7 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Lock "c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.819s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.651514] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.492s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.651943] env[61906]: INFO nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1] During sync_power_state the instance has a pending task (spawning). Skip. [ 517.653403] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "c6f7ccb5-f72d-4e97-89ed-f955dc58a5d1" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.689077] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquiring lock "f72e6538-7130-45a9-a722-ba60de3bda49" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.689271] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Lock "f72e6538-7130-45a9-a722-ba60de3bda49" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.783472] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356447, 'name': ReconfigVM_Task, 'duration_secs': 0.267169} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 517.783765] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45/85effb81-494c-4bda-a093-6dde523d5f45.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 517.785159] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8428a568-302f-45fc-8cb7-da9d53286b82 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.791272] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 517.791272] env[61906]: value = "task-1356448" [ 517.791272] env[61906]: _type = "Task" [ 517.791272] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 517.805291] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356448, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.117543] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.844s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 518.118662] env[61906]: ERROR nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2c453256-9d51-46a3-b814-54223f5fc641, please check neutron logs for more information. [ 518.118662] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Traceback (most recent call last): [ 518.118662] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 518.118662] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] self.driver.spawn(context, instance, image_meta, [ 518.118662] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 518.118662] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] self._vmops.spawn(context, instance, image_meta, injected_files, [ 518.118662] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 518.118662] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] vm_ref = self.build_virtual_machine(instance, [ 518.118662] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 518.118662] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] vif_infos = vmwarevif.get_vif_info(self._session, [ 518.118662] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 518.119111] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] for vif in network_info: [ 518.119111] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 518.119111] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] return self._sync_wrapper(fn, *args, **kwargs) [ 518.119111] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 518.119111] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] self.wait() [ 518.119111] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 518.119111] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] self[:] = self._gt.wait() [ 518.119111] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 518.119111] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] return self._exit_event.wait() [ 518.119111] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 518.119111] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] result = hub.switch() [ 518.119111] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 518.119111] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] return self.greenlet.switch() [ 518.119425] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 518.119425] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] result = function(*args, **kwargs) [ 518.119425] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 518.119425] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] return func(*args, **kwargs) [ 518.119425] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 518.119425] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] raise e [ 518.119425] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 518.119425] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] nwinfo = self.network_api.allocate_for_instance( [ 518.119425] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 518.119425] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] created_port_ids = self._update_ports_for_instance( [ 518.119425] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 518.119425] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] with excutils.save_and_reraise_exception(): [ 518.119425] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 518.120078] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] self.force_reraise() [ 518.120078] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 518.120078] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] raise self.value [ 518.120078] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 518.120078] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] updated_port = self._update_port( [ 518.120078] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 518.120078] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] _ensure_no_port_binding_failure(port) [ 518.120078] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 518.120078] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] raise exception.PortBindingFailed(port_id=port['id']) [ 518.120078] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] nova.exception.PortBindingFailed: Binding failed for port 2c453256-9d51-46a3-b814-54223f5fc641, please check neutron logs for more information. [ 518.120078] env[61906]: ERROR nova.compute.manager [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] [ 518.120436] env[61906]: DEBUG nova.compute.utils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Binding failed for port 2c453256-9d51-46a3-b814-54223f5fc641, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 518.120729] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.070s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.123066] env[61906]: INFO nova.compute.claims [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 518.126549] env[61906]: DEBUG nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Build of instance a3e97bd0-34e0-45f2-a10b-f554e10ca710 was re-scheduled: Binding failed for port 2c453256-9d51-46a3-b814-54223f5fc641, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 518.126549] env[61906]: DEBUG nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 518.126549] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Acquiring lock "refresh_cache-a3e97bd0-34e0-45f2-a10b-f554e10ca710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 518.126922] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Acquired lock "refresh_cache-a3e97bd0-34e0-45f2-a10b-f554e10ca710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 518.126922] env[61906]: DEBUG nova.network.neutron [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 518.140664] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e1f191b-f1ce-4ab7-a77d-2f35020cd004 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Lock "7f4bb9fb-fa64-451d-9876-7e4314c13a64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.468s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 518.143937] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "7f4bb9fb-fa64-451d-9876-7e4314c13a64" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.978s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.143937] env[61906]: INFO nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] During sync_power_state the instance has a pending task (block_device_mapping). Skip. [ 518.143937] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "7f4bb9fb-fa64-451d-9876-7e4314c13a64" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.002s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 518.154714] env[61906]: DEBUG nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.232169] env[61906]: DEBUG nova.network.neutron [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Successfully created port: 74e04542-2f68-486a-8d7c-48ae3f35e469 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 518.311182] env[61906]: DEBUG nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 518.313984] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356448, 'name': Rename_Task, 'duration_secs': 0.140624} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 518.314572] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 518.315176] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01b49f9f-d9f2-4f71-95a0-7c65b7923235 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.322505] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 518.322505] env[61906]: value = "task-1356449" [ 518.322505] env[61906]: _type = "Task" [ 518.322505] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 518.333704] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356449, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.356956] env[61906]: DEBUG nova.virt.hardware [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 518.357227] env[61906]: DEBUG nova.virt.hardware [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 518.357381] env[61906]: DEBUG nova.virt.hardware [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 518.357559] env[61906]: DEBUG nova.virt.hardware [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 518.357700] env[61906]: DEBUG nova.virt.hardware [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 518.357852] env[61906]: DEBUG nova.virt.hardware [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 518.358057] env[61906]: DEBUG nova.virt.hardware [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 518.358214] env[61906]: DEBUG nova.virt.hardware [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 518.358387] env[61906]: DEBUG nova.virt.hardware [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 518.358736] env[61906]: DEBUG nova.virt.hardware [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 518.358946] env[61906]: DEBUG nova.virt.hardware [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 518.359873] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e744c0-d830-4eed-8564-51d9c3ce8c67 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.367945] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf19bb62-d4b8-4996-a4b7-0084cf708efb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.435108] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Acquiring lock "6b3ecea4-fe20-4b6c-a987-59c20731b9ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.435613] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Lock "6b3ecea4-fe20-4b6c-a987-59c20731b9ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.647910] env[61906]: DEBUG nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.666223] env[61906]: DEBUG nova.network.neutron [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 518.688043] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.838757] env[61906]: DEBUG oslo_vmware.api [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356449, 'name': PowerOnVM_Task, 'duration_secs': 0.419711} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 518.839073] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 518.839289] env[61906]: DEBUG nova.compute.manager [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 518.840110] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ad148d-b995-4304-87d9-5a2c75826c07 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.945948] env[61906]: DEBUG nova.network.neutron [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 519.010133] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Acquiring lock "d23d0d19-4806-4d5e-aa05-13aaf3c4a770" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.010460] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Lock "d23d0d19-4806-4d5e-aa05-13aaf3c4a770" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.183711] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.363774] env[61906]: DEBUG oslo_concurrency.lockutils [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.437082] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb564b64-9183-42f3-8fc3-2a11abb3c48b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.445446] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2ef04c2-de76-4825-adfa-5c7e5f803a1a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.449173] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Releasing lock "refresh_cache-a3e97bd0-34e0-45f2-a10b-f554e10ca710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 519.449173] env[61906]: DEBUG nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 519.449341] env[61906]: DEBUG nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 519.449647] env[61906]: DEBUG nova.network.neutron [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 519.482319] env[61906]: DEBUG nova.network.neutron [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 519.484106] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdba312c-3e5b-492c-b71c-9d7be1beda51 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.492353] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933fcf9a-ef7c-40be-93c0-b19254928a9f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.508092] env[61906]: DEBUG nova.compute.provider_tree [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 519.991280] env[61906]: DEBUG nova.network.neutron [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.016431] env[61906]: DEBUG nova.scheduler.client.report [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 520.236135] env[61906]: ERROR nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 74e04542-2f68-486a-8d7c-48ae3f35e469, please check neutron logs for more information. [ 520.236135] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 520.236135] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 520.236135] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 520.236135] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 520.236135] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 520.236135] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 520.236135] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 520.236135] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.236135] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 520.236135] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.236135] env[61906]: ERROR nova.compute.manager raise self.value [ 520.236135] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 520.236135] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 520.236135] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.236135] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 520.236772] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.236772] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 520.236772] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 74e04542-2f68-486a-8d7c-48ae3f35e469, please check neutron logs for more information. [ 520.236772] env[61906]: ERROR nova.compute.manager [ 520.236772] env[61906]: Traceback (most recent call last): [ 520.236772] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 520.236772] env[61906]: listener.cb(fileno) [ 520.236772] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 520.236772] env[61906]: result = function(*args, **kwargs) [ 520.236772] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 520.236772] env[61906]: return func(*args, **kwargs) [ 520.236772] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 520.236772] env[61906]: raise e [ 520.236772] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 520.236772] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 520.236772] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 520.236772] env[61906]: created_port_ids = self._update_ports_for_instance( [ 520.236772] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 520.236772] env[61906]: with excutils.save_and_reraise_exception(): [ 520.236772] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.236772] env[61906]: self.force_reraise() [ 520.236772] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.236772] env[61906]: raise self.value [ 520.236772] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 520.236772] env[61906]: updated_port = self._update_port( [ 520.236772] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.236772] env[61906]: _ensure_no_port_binding_failure(port) [ 520.236772] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.236772] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 520.237495] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 74e04542-2f68-486a-8d7c-48ae3f35e469, please check neutron logs for more information. [ 520.237495] env[61906]: Removing descriptor: 17 [ 520.237495] env[61906]: ERROR nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 74e04542-2f68-486a-8d7c-48ae3f35e469, please check neutron logs for more information. [ 520.237495] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Traceback (most recent call last): [ 520.237495] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 520.237495] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] yield resources [ 520.237495] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 520.237495] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] self.driver.spawn(context, instance, image_meta, [ 520.237495] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 520.237495] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 520.237495] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 520.237495] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] vm_ref = self.build_virtual_machine(instance, [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] vif_infos = vmwarevif.get_vif_info(self._session, [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] for vif in network_info: [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] return self._sync_wrapper(fn, *args, **kwargs) [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] self.wait() [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] self[:] = self._gt.wait() [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] return self._exit_event.wait() [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 520.237832] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] result = hub.switch() [ 520.238186] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 520.238186] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] return self.greenlet.switch() [ 520.238186] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 520.238186] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] result = function(*args, **kwargs) [ 520.238186] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 520.238186] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] return func(*args, **kwargs) [ 520.238186] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 520.238186] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] raise e [ 520.238186] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 520.238186] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] nwinfo = self.network_api.allocate_for_instance( [ 520.238186] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 520.238186] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] created_port_ids = self._update_ports_for_instance( [ 520.238186] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 520.238532] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] with excutils.save_and_reraise_exception(): [ 520.238532] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.238532] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] self.force_reraise() [ 520.238532] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.238532] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] raise self.value [ 520.238532] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 520.238532] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] updated_port = self._update_port( [ 520.238532] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.238532] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] _ensure_no_port_binding_failure(port) [ 520.238532] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.238532] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] raise exception.PortBindingFailed(port_id=port['id']) [ 520.238532] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] nova.exception.PortBindingFailed: Binding failed for port 74e04542-2f68-486a-8d7c-48ae3f35e469, please check neutron logs for more information. [ 520.238532] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] [ 520.238909] env[61906]: INFO nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Terminating instance [ 520.240122] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Acquiring lock "refresh_cache-fa0cbe6c-937b-4276-a419-0ab6e92e795f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 520.240122] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Acquired lock "refresh_cache-fa0cbe6c-937b-4276-a419-0ab6e92e795f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 520.240122] env[61906]: DEBUG nova.network.neutron [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 520.494717] env[61906]: INFO nova.compute.manager [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] Took 1.04 seconds to deallocate network for instance. [ 520.517688] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.518211] env[61906]: DEBUG nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 520.522909] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 7.834s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.529017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.003s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.529017] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 520.529017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.495s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.529017] env[61906]: INFO nova.compute.claims [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 520.531714] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c44f1e-4e32-4205-a82a-8905e3643c5c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.543145] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5608b5c1-0821-45e2-b076-96b0fcbba8a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.549540] env[61906]: DEBUG nova.compute.manager [None req-868cd0f1-5a11-4749-82f0-492115d5c33f tempest-ServerDiagnosticsV248Test-1288319871 tempest-ServerDiagnosticsV248Test-1288319871-project-admin] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 520.552761] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effc304d-caf7-4ba6-ac34-dd5af097a207 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.567212] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30cd382f-ea09-4b5d-959e-a87de70d69fd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.573689] env[61906]: INFO nova.compute.manager [None req-868cd0f1-5a11-4749-82f0-492115d5c33f tempest-ServerDiagnosticsV248Test-1288319871 tempest-ServerDiagnosticsV248Test-1288319871-project-admin] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Retrieving diagnostics [ 520.575101] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0213cc2-21eb-4f30-95c7-c5a10c03f421 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.581069] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c0020a-9622-4d04-9ebd-27b5be2c9b63 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.620519] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Acquiring lock "528a2095-e82e-44eb-a4a1-dacde88966bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.620818] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Lock "528a2095-e82e-44eb-a4a1-dacde88966bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.645736] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181503MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 520.645943] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.768822] env[61906]: DEBUG nova.network.neutron [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 520.841049] env[61906]: INFO nova.compute.manager [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Rebuilding instance [ 520.882496] env[61906]: DEBUG nova.network.neutron [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.908457] env[61906]: DEBUG nova.compute.manager [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 520.912047] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-802a3c38-bbc4-49eb-9b87-6d2892b1f736 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.034060] env[61906]: DEBUG nova.compute.utils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 521.035814] env[61906]: DEBUG nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 521.035922] env[61906]: DEBUG nova.network.neutron [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 521.206018] env[61906]: DEBUG nova.policy [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '303dd21388ad47e69e5d32edbeeb6804', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e28297ad5784e82a6d43cc6ef86d690', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 521.388022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Releasing lock "refresh_cache-fa0cbe6c-937b-4276-a419-0ab6e92e795f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.388022] env[61906]: DEBUG nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 521.388022] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 521.388022] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3353c42-e5f0-4964-a2d3-b26146c31621 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.398822] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e79e8c-76c9-46a1-a8ba-88612a528950 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.425037] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 521.431711] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6aa3a06-346d-4b25-aa80-f3a63feb3e59 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.434901] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fa0cbe6c-937b-4276-a419-0ab6e92e795f could not be found. [ 521.435503] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 521.436048] env[61906]: INFO nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 521.437271] env[61906]: DEBUG oslo.service.loopingcall [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 521.437271] env[61906]: DEBUG nova.compute.manager [-] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 521.437271] env[61906]: DEBUG nova.network.neutron [-] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 521.446700] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Waiting for the task: (returnval){ [ 521.446700] env[61906]: value = "task-1356454" [ 521.446700] env[61906]: _type = "Task" [ 521.446700] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.459209] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356454, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.466701] env[61906]: DEBUG nova.network.neutron [-] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 521.544660] env[61906]: DEBUG nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 521.556020] env[61906]: INFO nova.scheduler.client.report [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Deleted allocations for instance a3e97bd0-34e0-45f2-a10b-f554e10ca710 [ 521.760881] env[61906]: DEBUG nova.network.neutron [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Successfully created port: 65707ace-cf18-4ce6-8f31-6802abce3c29 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 521.800559] env[61906]: DEBUG nova.compute.manager [req-ee0e8233-10f2-4c1c-a5d4-d5eaa55658b2 req-dfc29d35-d037-4843-9881-d5439e5de311 service nova] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Received event network-changed-74e04542-2f68-486a-8d7c-48ae3f35e469 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 521.800674] env[61906]: DEBUG nova.compute.manager [req-ee0e8233-10f2-4c1c-a5d4-d5eaa55658b2 req-dfc29d35-d037-4843-9881-d5439e5de311 service nova] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Refreshing instance network info cache due to event network-changed-74e04542-2f68-486a-8d7c-48ae3f35e469. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 521.800846] env[61906]: DEBUG oslo_concurrency.lockutils [req-ee0e8233-10f2-4c1c-a5d4-d5eaa55658b2 req-dfc29d35-d037-4843-9881-d5439e5de311 service nova] Acquiring lock "refresh_cache-fa0cbe6c-937b-4276-a419-0ab6e92e795f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.801011] env[61906]: DEBUG oslo_concurrency.lockutils [req-ee0e8233-10f2-4c1c-a5d4-d5eaa55658b2 req-dfc29d35-d037-4843-9881-d5439e5de311 service nova] Acquired lock "refresh_cache-fa0cbe6c-937b-4276-a419-0ab6e92e795f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.801206] env[61906]: DEBUG nova.network.neutron [req-ee0e8233-10f2-4c1c-a5d4-d5eaa55658b2 req-dfc29d35-d037-4843-9881-d5439e5de311 service nova] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Refreshing network info cache for port 74e04542-2f68-486a-8d7c-48ae3f35e469 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 521.847058] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e204abcb-868f-4910-86fb-b6f42bf84f49 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.860067] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610fc0ba-2287-4bf7-998b-fab21eebf060 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.894303] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71131d9-957c-4a2c-8ebc-b0528475f69f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.902048] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10900a7e-e694-467c-868f-162d6568963e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.916517] env[61906]: DEBUG nova.compute.provider_tree [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.957910] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356454, 'name': PowerOffVM_Task, 'duration_secs': 0.255409} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.958028] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 521.958164] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 521.959103] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39642cf7-6a36-4573-a514-addeb6762c82 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.966804] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 521.966804] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-423c1245-f4c8-4dfe-bbe8-64cbe39f531d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.968412] env[61906]: DEBUG nova.network.neutron [-] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 521.997762] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 521.998258] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 522.000281] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Deleting the datastore file [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 522.000281] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-407dcf34-f2fd-4682-abec-6a7f0724b357 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.007187] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Waiting for the task: (returnval){ [ 522.007187] env[61906]: value = "task-1356457" [ 522.007187] env[61906]: _type = "Task" [ 522.007187] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.019349] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356457, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.069403] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4055482f-3ab5-4fc1-b53f-89483fc6d7a8 tempest-ImagesOneServerTestJSON-689347432 tempest-ImagesOneServerTestJSON-689347432-project-member] Lock "a3e97bd0-34e0-45f2-a10b-f554e10ca710" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.163s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.070966] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "a3e97bd0-34e0-45f2-a10b-f554e10ca710" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 9.911s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.071279] env[61906]: INFO nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: a3e97bd0-34e0-45f2-a10b-f554e10ca710] During sync_power_state the instance has a pending task (spawning). Skip. [ 522.071430] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "a3e97bd0-34e0-45f2-a10b-f554e10ca710" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.328768] env[61906]: DEBUG nova.network.neutron [req-ee0e8233-10f2-4c1c-a5d4-d5eaa55658b2 req-dfc29d35-d037-4843-9881-d5439e5de311 service nova] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 522.421782] env[61906]: DEBUG nova.scheduler.client.report [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 522.461588] env[61906]: DEBUG nova.network.neutron [req-ee0e8233-10f2-4c1c-a5d4-d5eaa55658b2 req-dfc29d35-d037-4843-9881-d5439e5de311 service nova] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.471271] env[61906]: INFO nova.compute.manager [-] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Took 1.03 seconds to deallocate network for instance. [ 522.476482] env[61906]: DEBUG nova.compute.claims [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 522.476482] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.521728] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356457, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173599} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.522034] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 522.522239] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 522.522416] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 522.563125] env[61906]: DEBUG nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 522.573106] env[61906]: DEBUG nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 522.598703] env[61906]: DEBUG nova.virt.hardware [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:04:44Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='2091665628',id=21,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-683227710',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 522.598703] env[61906]: DEBUG nova.virt.hardware [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 522.598703] env[61906]: DEBUG nova.virt.hardware [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 522.598912] env[61906]: DEBUG nova.virt.hardware [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 522.598912] env[61906]: DEBUG nova.virt.hardware [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 522.598912] env[61906]: DEBUG nova.virt.hardware [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 522.598912] env[61906]: DEBUG nova.virt.hardware [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 522.598912] env[61906]: DEBUG nova.virt.hardware [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 522.599283] env[61906]: DEBUG nova.virt.hardware [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 522.599283] env[61906]: DEBUG nova.virt.hardware [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 522.599283] env[61906]: DEBUG nova.virt.hardware [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 522.599692] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-344bc7e6-05bb-4c9f-8ea0-8df2f411c592 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.614328] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa09a694-84be-48ac-a7d4-72760389ad38 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.928219] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.928773] env[61906]: DEBUG nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 522.935507] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.739s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.968354] env[61906]: DEBUG oslo_concurrency.lockutils [req-ee0e8233-10f2-4c1c-a5d4-d5eaa55658b2 req-dfc29d35-d037-4843-9881-d5439e5de311 service nova] Releasing lock "refresh_cache-fa0cbe6c-937b-4276-a419-0ab6e92e795f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.108113] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.408710] env[61906]: ERROR nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 65707ace-cf18-4ce6-8f31-6802abce3c29, please check neutron logs for more information. [ 523.408710] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 523.408710] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.408710] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 523.408710] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.408710] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 523.408710] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.408710] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 523.408710] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.408710] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 523.408710] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.408710] env[61906]: ERROR nova.compute.manager raise self.value [ 523.408710] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.408710] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 523.408710] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.408710] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 523.409309] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.409309] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 523.409309] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 65707ace-cf18-4ce6-8f31-6802abce3c29, please check neutron logs for more information. [ 523.409309] env[61906]: ERROR nova.compute.manager [ 523.409309] env[61906]: Traceback (most recent call last): [ 523.409309] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 523.409309] env[61906]: listener.cb(fileno) [ 523.409309] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.409309] env[61906]: result = function(*args, **kwargs) [ 523.409309] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 523.409309] env[61906]: return func(*args, **kwargs) [ 523.409309] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.409309] env[61906]: raise e [ 523.409309] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.409309] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 523.409309] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.409309] env[61906]: created_port_ids = self._update_ports_for_instance( [ 523.409309] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.409309] env[61906]: with excutils.save_and_reraise_exception(): [ 523.409309] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.409309] env[61906]: self.force_reraise() [ 523.409309] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.409309] env[61906]: raise self.value [ 523.409309] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.409309] env[61906]: updated_port = self._update_port( [ 523.409309] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.409309] env[61906]: _ensure_no_port_binding_failure(port) [ 523.409309] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.409309] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 523.410140] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 65707ace-cf18-4ce6-8f31-6802abce3c29, please check neutron logs for more information. [ 523.410140] env[61906]: Removing descriptor: 17 [ 523.410140] env[61906]: ERROR nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 65707ace-cf18-4ce6-8f31-6802abce3c29, please check neutron logs for more information. [ 523.410140] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Traceback (most recent call last): [ 523.410140] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 523.410140] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] yield resources [ 523.410140] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.410140] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] self.driver.spawn(context, instance, image_meta, [ 523.410140] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 523.410140] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.410140] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.410140] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] vm_ref = self.build_virtual_machine(instance, [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] for vif in network_info: [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] return self._sync_wrapper(fn, *args, **kwargs) [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] self.wait() [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] self[:] = self._gt.wait() [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] return self._exit_event.wait() [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.410549] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] result = hub.switch() [ 523.410948] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.410948] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] return self.greenlet.switch() [ 523.410948] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.410948] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] result = function(*args, **kwargs) [ 523.410948] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 523.410948] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] return func(*args, **kwargs) [ 523.410948] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.410948] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] raise e [ 523.410948] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.410948] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] nwinfo = self.network_api.allocate_for_instance( [ 523.410948] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.410948] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] created_port_ids = self._update_ports_for_instance( [ 523.410948] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.411377] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] with excutils.save_and_reraise_exception(): [ 523.411377] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.411377] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] self.force_reraise() [ 523.411377] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.411377] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] raise self.value [ 523.411377] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.411377] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] updated_port = self._update_port( [ 523.411377] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.411377] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] _ensure_no_port_binding_failure(port) [ 523.411377] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.411377] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] raise exception.PortBindingFailed(port_id=port['id']) [ 523.411377] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] nova.exception.PortBindingFailed: Binding failed for port 65707ace-cf18-4ce6-8f31-6802abce3c29, please check neutron logs for more information. [ 523.411377] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] [ 523.411754] env[61906]: INFO nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Terminating instance [ 523.412108] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Acquiring lock "refresh_cache-9f6864a2-3e54-41cb-b858-16e6e350a2af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.412272] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Acquired lock "refresh_cache-9f6864a2-3e54-41cb-b858-16e6e350a2af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.412435] env[61906]: DEBUG nova.network.neutron [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 523.442890] env[61906]: DEBUG nova.compute.utils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.451919] env[61906]: DEBUG nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 523.451919] env[61906]: DEBUG nova.network.neutron [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 523.530058] env[61906]: DEBUG nova.policy [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '92b33b58938746279209ebc8d749f604', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e373eb0dc4d47bbb200e6cbe8c2c661', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 523.562231] env[61906]: DEBUG nova.virt.hardware [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 523.562507] env[61906]: DEBUG nova.virt.hardware [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 523.562682] env[61906]: DEBUG nova.virt.hardware [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 523.562842] env[61906]: DEBUG nova.virt.hardware [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 523.563082] env[61906]: DEBUG nova.virt.hardware [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 523.563264] env[61906]: DEBUG nova.virt.hardware [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 523.563490] env[61906]: DEBUG nova.virt.hardware [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 523.563623] env[61906]: DEBUG nova.virt.hardware [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 523.563856] env[61906]: DEBUG nova.virt.hardware [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 523.564053] env[61906]: DEBUG nova.virt.hardware [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 523.564238] env[61906]: DEBUG nova.virt.hardware [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 523.565154] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87569656-cc5d-4998-a627-8e1bf46c1b93 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.576447] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6af80f-aca9-4667-adb6-5f6784bd7aaf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.591475] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 523.596895] env[61906]: DEBUG oslo.service.loopingcall [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 523.601584] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 523.602194] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d2e366e1-8dd3-4480-9236-fd9d7ee59774 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.620112] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 523.620112] env[61906]: value = "task-1356458" [ 523.620112] env[61906]: _type = "Task" [ 523.620112] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.627413] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356458, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.775709] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27c7e39-e696-4448-8c39-c6fe0628296b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.783864] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f475e0e8-c915-4fc7-a189-ab345088fb6b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.826739] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5ed30d-6605-4ca2-9b3d-fe8f1cadbf19 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.837171] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7538d864-f5dc-44ce-b709-e6d9717edcad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.853371] env[61906]: DEBUG nova.compute.provider_tree [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 523.928674] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquiring lock "e64fd277-10ae-494d-918a-2f3faa4e44d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.928905] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Lock "e64fd277-10ae-494d-918a-2f3faa4e44d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.941419] env[61906]: DEBUG nova.network.neutron [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.950129] env[61906]: DEBUG nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 524.084604] env[61906]: DEBUG nova.network.neutron [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Successfully created port: 7caa7da5-777c-43bf-8628-f8741f621e94 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 524.129854] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356458, 'name': CreateVM_Task, 'duration_secs': 0.303354} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.130808] env[61906]: DEBUG nova.network.neutron [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.132021] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 524.132412] env[61906]: DEBUG oslo_concurrency.lockutils [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.132567] env[61906]: DEBUG oslo_concurrency.lockutils [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.132875] env[61906]: DEBUG oslo_concurrency.lockutils [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 524.133580] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c0a651f-8ca7-427f-bf98-e83ab8436c5f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.141707] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Waiting for the task: (returnval){ [ 524.141707] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b80a3f-5d03-f30d-f30b-223ea8d47cce" [ 524.141707] env[61906]: _type = "Task" [ 524.141707] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.158939] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b80a3f-5d03-f30d-f30b-223ea8d47cce, 'name': SearchDatastore_Task, 'duration_secs': 0.01098} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.159260] env[61906]: DEBUG oslo_concurrency.lockutils [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.159481] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 524.159707] env[61906]: DEBUG oslo_concurrency.lockutils [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.159962] env[61906]: DEBUG oslo_concurrency.lockutils [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.160048] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 524.160286] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aad93e18-6d0c-44d0-8562-e0fce6bd9136 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.170126] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 524.170330] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 524.171048] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-199f5429-39af-4eb5-9c41-ee9727752f47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.184428] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Waiting for the task: (returnval){ [ 524.184428] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52579701-cd0f-b855-5847-52c3eed39c9a" [ 524.184428] env[61906]: _type = "Task" [ 524.184428] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.194710] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52579701-cd0f-b855-5847-52c3eed39c9a, 'name': SearchDatastore_Task, 'duration_secs': 0.009192} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.196034] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-085c1ccb-8d47-4b03-975a-cc019f1a2367 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.202261] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Waiting for the task: (returnval){ [ 524.202261] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527e357d-87f9-9015-f935-e2c7f2e42447" [ 524.202261] env[61906]: _type = "Task" [ 524.202261] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.211561] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527e357d-87f9-9015-f935-e2c7f2e42447, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.288009] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Acquiring lock "4b0abc14-0ce2-4212-8677-82f6279ee107" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.288263] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Lock "4b0abc14-0ce2-4212-8677-82f6279ee107" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.376905] env[61906]: ERROR nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [req-456449aa-e76d-4a45-a0eb-a19a5c49540d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6e524f5d-985e-41eb-a7ae-996be3ae1c1a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-456449aa-e76d-4a45-a0eb-a19a5c49540d"}]}: nova.exception.PortBindingFailed: Binding failed for port 5fbe7162-22fd-463d-a62c-1725806badc1, please check neutron logs for more information. [ 524.402775] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Refreshing inventories for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 524.418896] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Updating ProviderTree inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 524.418896] env[61906]: DEBUG nova.compute.provider_tree [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 524.435018] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Refreshing aggregate associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 524.465713] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Refreshing trait associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 524.634134] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Releasing lock "refresh_cache-9f6864a2-3e54-41cb-b858-16e6e350a2af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.634555] env[61906]: DEBUG nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 524.634748] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 524.637801] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-090d3ff7-bf2e-4faa-8a26-48e2faead633 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.647062] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a1c70e-5629-4d46-a386-191378803b3e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.675537] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9f6864a2-3e54-41cb-b858-16e6e350a2af could not be found. [ 524.675778] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 524.675965] env[61906]: INFO nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Took 0.04 seconds to destroy the instance on the hypervisor. [ 524.676312] env[61906]: DEBUG oslo.service.loopingcall [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 524.679607] env[61906]: DEBUG nova.compute.manager [-] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 524.679749] env[61906]: DEBUG nova.network.neutron [-] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 524.710069] env[61906]: DEBUG nova.network.neutron [-] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.716416] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527e357d-87f9-9015-f935-e2c7f2e42447, 'name': SearchDatastore_Task, 'duration_secs': 0.009239} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.720951] env[61906]: DEBUG oslo_concurrency.lockutils [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.722029] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45/85effb81-494c-4bda-a093-6dde523d5f45.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 524.724177] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8e04dff8-0669-45d0-9abf-a459464c5ae3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.734248] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Waiting for the task: (returnval){ [ 524.734248] env[61906]: value = "task-1356460" [ 524.734248] env[61906]: _type = "Task" [ 524.734248] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.748934] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356460, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.777288] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68df947-b6f2-449e-9ed6-268bacb8e422 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.788412] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e937da40-40b1-4d46-8243-c9f6e0fa569e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.826601] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f39e0c08-e697-4639-9308-61cb0c18f6e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.835603] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7831b86a-371d-4a08-b05d-1bc07f817abd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.854509] env[61906]: DEBUG nova.compute.provider_tree [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 524.960819] env[61906]: DEBUG nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 525.000274] env[61906]: DEBUG nova.virt.hardware [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 525.000274] env[61906]: DEBUG nova.virt.hardware [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 525.000274] env[61906]: DEBUG nova.virt.hardware [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 525.000451] env[61906]: DEBUG nova.virt.hardware [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 525.000451] env[61906]: DEBUG nova.virt.hardware [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 525.000451] env[61906]: DEBUG nova.virt.hardware [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 525.000451] env[61906]: DEBUG nova.virt.hardware [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 525.000451] env[61906]: DEBUG nova.virt.hardware [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 525.000676] env[61906]: DEBUG nova.virt.hardware [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 525.000676] env[61906]: DEBUG nova.virt.hardware [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 525.000676] env[61906]: DEBUG nova.virt.hardware [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 525.001820] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9b91b7-b8d1-46cc-ac79-7db379e6353c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.014395] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0208ef7e-95db-4d6f-92c1-41e2722a7c4f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.072476] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Acquiring lock "05cefc82-3585-4d0b-8620-91412b3ef69e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.073218] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Lock "05cefc82-3585-4d0b-8620-91412b3ef69e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.098886] env[61906]: DEBUG nova.compute.manager [req-a9cb70df-0c89-4203-97f9-0e6c9c6b5361 req-eb64cd12-3721-4959-9711-d6757f85043b service nova] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Received event network-vif-deleted-74e04542-2f68-486a-8d7c-48ae3f35e469 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 525.100458] env[61906]: DEBUG nova.compute.manager [req-a9cb70df-0c89-4203-97f9-0e6c9c6b5361 req-eb64cd12-3721-4959-9711-d6757f85043b service nova] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Received event network-changed-65707ace-cf18-4ce6-8f31-6802abce3c29 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 525.100774] env[61906]: DEBUG nova.compute.manager [req-a9cb70df-0c89-4203-97f9-0e6c9c6b5361 req-eb64cd12-3721-4959-9711-d6757f85043b service nova] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Refreshing instance network info cache due to event network-changed-65707ace-cf18-4ce6-8f31-6802abce3c29. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 525.101127] env[61906]: DEBUG oslo_concurrency.lockutils [req-a9cb70df-0c89-4203-97f9-0e6c9c6b5361 req-eb64cd12-3721-4959-9711-d6757f85043b service nova] Acquiring lock "refresh_cache-9f6864a2-3e54-41cb-b858-16e6e350a2af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.101498] env[61906]: DEBUG oslo_concurrency.lockutils [req-a9cb70df-0c89-4203-97f9-0e6c9c6b5361 req-eb64cd12-3721-4959-9711-d6757f85043b service nova] Acquired lock "refresh_cache-9f6864a2-3e54-41cb-b858-16e6e350a2af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.101573] env[61906]: DEBUG nova.network.neutron [req-a9cb70df-0c89-4203-97f9-0e6c9c6b5361 req-eb64cd12-3721-4959-9711-d6757f85043b service nova] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Refreshing network info cache for port 65707ace-cf18-4ce6-8f31-6802abce3c29 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 525.220988] env[61906]: DEBUG nova.network.neutron [-] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.250483] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356460, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.382675] env[61906]: ERROR nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [req-9d597cf3-c5ef-4fca-b6f4-5dace06465a4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6e524f5d-985e-41eb-a7ae-996be3ae1c1a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9d597cf3-c5ef-4fca-b6f4-5dace06465a4"}]}: nova.exception.PortBindingFailed: Binding failed for port 5fbe7162-22fd-463d-a62c-1725806badc1, please check neutron logs for more information. [ 525.403476] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Refreshing inventories for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 525.420216] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Updating ProviderTree inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 525.420448] env[61906]: DEBUG nova.compute.provider_tree [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 525.438301] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Refreshing aggregate associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 525.459022] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Refreshing trait associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 525.649044] env[61906]: DEBUG nova.network.neutron [req-a9cb70df-0c89-4203-97f9-0e6c9c6b5361 req-eb64cd12-3721-4959-9711-d6757f85043b service nova] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 525.726030] env[61906]: INFO nova.compute.manager [-] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Took 1.05 seconds to deallocate network for instance. [ 525.728768] env[61906]: DEBUG nova.compute.claims [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 525.728980] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.751952] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356460, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534671} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.753153] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45/85effb81-494c-4bda-a093-6dde523d5f45.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 525.753254] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 525.753947] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b20319-d36e-49b2-9c60-d16fad5dad3d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.756581] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f90d5aa3-dd19-4aa8-8f4d-f8869c64a760 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.765008] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd637019-5c0c-422c-a3f1-a5cb9b7bccd7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.771275] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Waiting for the task: (returnval){ [ 525.771275] env[61906]: value = "task-1356461" [ 525.771275] env[61906]: _type = "Task" [ 525.771275] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.806709] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac9e2d9-3973-4905-853f-1ca44889a41f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.813709] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356461, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.818620] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e42810-4602-4892-9a1c-18ad94c7eb46 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.835812] env[61906]: DEBUG nova.compute.provider_tree [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 525.931213] env[61906]: DEBUG nova.network.neutron [req-a9cb70df-0c89-4203-97f9-0e6c9c6b5361 req-eb64cd12-3721-4959-9711-d6757f85043b service nova] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.283331] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356461, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065439} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.283331] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 526.284013] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1eed983-c68a-48f2-970b-1621aba931bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.306125] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45/85effb81-494c-4bda-a093-6dde523d5f45.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 526.306473] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c53c609-57e4-46f7-8e45-b7c9d7a4d3f3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.325557] env[61906]: ERROR nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7caa7da5-777c-43bf-8628-f8741f621e94, please check neutron logs for more information. [ 526.325557] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 526.325557] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.325557] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 526.325557] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.325557] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 526.325557] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.325557] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 526.325557] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.325557] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 526.325557] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.325557] env[61906]: ERROR nova.compute.manager raise self.value [ 526.325557] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.325557] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 526.325557] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.325557] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 526.326321] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.326321] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 526.326321] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7caa7da5-777c-43bf-8628-f8741f621e94, please check neutron logs for more information. [ 526.326321] env[61906]: ERROR nova.compute.manager [ 526.326321] env[61906]: Traceback (most recent call last): [ 526.326321] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 526.326321] env[61906]: listener.cb(fileno) [ 526.326321] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.326321] env[61906]: result = function(*args, **kwargs) [ 526.326321] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 526.326321] env[61906]: return func(*args, **kwargs) [ 526.326321] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 526.326321] env[61906]: raise e [ 526.326321] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.326321] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 526.326321] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.326321] env[61906]: created_port_ids = self._update_ports_for_instance( [ 526.326321] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.326321] env[61906]: with excutils.save_and_reraise_exception(): [ 526.326321] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.326321] env[61906]: self.force_reraise() [ 526.326321] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.326321] env[61906]: raise self.value [ 526.326321] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.326321] env[61906]: updated_port = self._update_port( [ 526.326321] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.326321] env[61906]: _ensure_no_port_binding_failure(port) [ 526.326321] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.326321] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 526.328466] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 7caa7da5-777c-43bf-8628-f8741f621e94, please check neutron logs for more information. [ 526.328466] env[61906]: Removing descriptor: 17 [ 526.328466] env[61906]: ERROR nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7caa7da5-777c-43bf-8628-f8741f621e94, please check neutron logs for more information. [ 526.328466] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Traceback (most recent call last): [ 526.328466] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 526.328466] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] yield resources [ 526.328466] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 526.328466] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] self.driver.spawn(context, instance, image_meta, [ 526.328466] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 526.328466] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] self._vmops.spawn(context, instance, image_meta, injected_files, [ 526.328466] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 526.328466] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] vm_ref = self.build_virtual_machine(instance, [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] vif_infos = vmwarevif.get_vif_info(self._session, [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] for vif in network_info: [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] return self._sync_wrapper(fn, *args, **kwargs) [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] self.wait() [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] self[:] = self._gt.wait() [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] return self._exit_event.wait() [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 526.328960] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] result = hub.switch() [ 526.330440] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 526.330440] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] return self.greenlet.switch() [ 526.330440] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.330440] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] result = function(*args, **kwargs) [ 526.330440] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 526.330440] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] return func(*args, **kwargs) [ 526.330440] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 526.330440] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] raise e [ 526.330440] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.330440] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] nwinfo = self.network_api.allocate_for_instance( [ 526.330440] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.330440] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] created_port_ids = self._update_ports_for_instance( [ 526.330440] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.331354] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] with excutils.save_and_reraise_exception(): [ 526.331354] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.331354] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] self.force_reraise() [ 526.331354] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.331354] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] raise self.value [ 526.331354] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.331354] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] updated_port = self._update_port( [ 526.331354] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.331354] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] _ensure_no_port_binding_failure(port) [ 526.331354] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.331354] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] raise exception.PortBindingFailed(port_id=port['id']) [ 526.331354] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] nova.exception.PortBindingFailed: Binding failed for port 7caa7da5-777c-43bf-8628-f8741f621e94, please check neutron logs for more information. [ 526.331354] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] [ 526.331983] env[61906]: INFO nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Terminating instance [ 526.331983] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Acquiring lock "refresh_cache-4dc3af05-b670-4f7c-bb9f-d1fa2e4180db" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.331983] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Acquired lock "refresh_cache-4dc3af05-b670-4f7c-bb9f-d1fa2e4180db" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.331983] env[61906]: DEBUG nova.network.neutron [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 526.340955] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Waiting for the task: (returnval){ [ 526.340955] env[61906]: value = "task-1356463" [ 526.340955] env[61906]: _type = "Task" [ 526.340955] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.352615] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356463, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.359600] env[61906]: ERROR nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [req-ad63507d-c978-4adf-9973-84225e419933] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6e524f5d-985e-41eb-a7ae-996be3ae1c1a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ad63507d-c978-4adf-9973-84225e419933"}]}: nova.exception.PortBindingFailed: Binding failed for port 5fbe7162-22fd-463d-a62c-1725806badc1, please check neutron logs for more information. [ 526.378320] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Refreshing inventories for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 526.396018] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Updating ProviderTree inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 526.396267] env[61906]: DEBUG nova.compute.provider_tree [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 526.408786] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Refreshing aggregate associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, aggregates: 5f28d5a7-b640-4216-a3f2-95116ad8b41b {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 526.429943] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Refreshing trait associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 526.434008] env[61906]: DEBUG oslo_concurrency.lockutils [req-a9cb70df-0c89-4203-97f9-0e6c9c6b5361 req-eb64cd12-3721-4959-9711-d6757f85043b service nova] Releasing lock "refresh_cache-9f6864a2-3e54-41cb-b858-16e6e350a2af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.434244] env[61906]: DEBUG nova.compute.manager [req-a9cb70df-0c89-4203-97f9-0e6c9c6b5361 req-eb64cd12-3721-4959-9711-d6757f85043b service nova] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Received event network-vif-deleted-65707ace-cf18-4ce6-8f31-6802abce3c29 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 526.798205] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde8af49-71ce-409d-8bb9-2bf8533a2eaa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.808263] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-971a9082-c023-41e6-ac94-75ceb47156c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.851539] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f176c3dc-8ac9-481f-81e5-b9bd43c041fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.865825] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66976f2-4c6f-44a5-929f-aab36ab95ce8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.871045] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356463, 'name': ReconfigVM_Task, 'duration_secs': 0.322126} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.871431] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45/85effb81-494c-4bda-a093-6dde523d5f45.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 526.872383] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3aa38461-b95c-4e83-b5f7-660851255a55 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.883594] env[61906]: DEBUG nova.compute.provider_tree [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 526.888586] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Waiting for the task: (returnval){ [ 526.888586] env[61906]: value = "task-1356464" [ 526.888586] env[61906]: _type = "Task" [ 526.888586] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.897170] env[61906]: DEBUG nova.network.neutron [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.903774] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356464, 'name': Rename_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.249486] env[61906]: DEBUG nova.network.neutron [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.409738] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356464, 'name': Rename_Task, 'duration_secs': 0.157731} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.409738] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 527.409738] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1228eadc-d989-4d13-81e2-cfe7405f5e4e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.415732] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Waiting for the task: (returnval){ [ 527.415732] env[61906]: value = "task-1356465" [ 527.415732] env[61906]: _type = "Task" [ 527.415732] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.426134] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356465, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.444435] env[61906]: DEBUG nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Updated inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with generation 27 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 527.444885] env[61906]: DEBUG nova.compute.provider_tree [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Updating resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a generation from 27 to 28 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 527.444885] env[61906]: DEBUG nova.compute.provider_tree [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 527.727924] env[61906]: DEBUG nova.compute.manager [req-3b8ed611-b751-4fb9-a304-7b5fabcf24b5 req-f7726de2-f45c-43a5-8291-9781ccbc4389 service nova] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Received event network-changed-7caa7da5-777c-43bf-8628-f8741f621e94 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.728167] env[61906]: DEBUG nova.compute.manager [req-3b8ed611-b751-4fb9-a304-7b5fabcf24b5 req-f7726de2-f45c-43a5-8291-9781ccbc4389 service nova] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Refreshing instance network info cache due to event network-changed-7caa7da5-777c-43bf-8628-f8741f621e94. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 527.728371] env[61906]: DEBUG oslo_concurrency.lockutils [req-3b8ed611-b751-4fb9-a304-7b5fabcf24b5 req-f7726de2-f45c-43a5-8291-9781ccbc4389 service nova] Acquiring lock "refresh_cache-4dc3af05-b670-4f7c-bb9f-d1fa2e4180db" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.754022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Releasing lock "refresh_cache-4dc3af05-b670-4f7c-bb9f-d1fa2e4180db" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.754022] env[61906]: DEBUG nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 527.754022] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 527.754022] env[61906]: DEBUG oslo_concurrency.lockutils [req-3b8ed611-b751-4fb9-a304-7b5fabcf24b5 req-f7726de2-f45c-43a5-8291-9781ccbc4389 service nova] Acquired lock "refresh_cache-4dc3af05-b670-4f7c-bb9f-d1fa2e4180db" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.754231] env[61906]: DEBUG nova.network.neutron [req-3b8ed611-b751-4fb9-a304-7b5fabcf24b5 req-f7726de2-f45c-43a5-8291-9781ccbc4389 service nova] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Refreshing network info cache for port 7caa7da5-777c-43bf-8628-f8741f621e94 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 527.755239] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4583d176-a70d-4a10-85fd-23d47a2d9689 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.775030] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f472bbb8-dc01-4180-9c96-bfe23874e097 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.808040] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db could not be found. [ 527.808277] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 527.808461] env[61906]: INFO nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Took 0.05 seconds to destroy the instance on the hypervisor. [ 527.808707] env[61906]: DEBUG oslo.service.loopingcall [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 527.809480] env[61906]: DEBUG nova.compute.manager [-] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 527.809480] env[61906]: DEBUG nova.network.neutron [-] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 527.846812] env[61906]: DEBUG nova.network.neutron [-] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.933528] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356465, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.950620] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 5.015s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.951278] env[61906]: ERROR nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5fbe7162-22fd-463d-a62c-1725806badc1, please check neutron logs for more information. [ 527.951278] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Traceback (most recent call last): [ 527.951278] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 527.951278] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] self.driver.spawn(context, instance, image_meta, [ 527.951278] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 527.951278] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 527.951278] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 527.951278] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] vm_ref = self.build_virtual_machine(instance, [ 527.951278] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 527.951278] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] vif_infos = vmwarevif.get_vif_info(self._session, [ 527.951278] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 527.951584] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] for vif in network_info: [ 527.951584] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 527.951584] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] return self._sync_wrapper(fn, *args, **kwargs) [ 527.951584] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 527.951584] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] self.wait() [ 527.951584] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 527.951584] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] self[:] = self._gt.wait() [ 527.951584] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 527.951584] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] return self._exit_event.wait() [ 527.951584] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 527.951584] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] result = hub.switch() [ 527.951584] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 527.951584] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] return self.greenlet.switch() [ 527.951934] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.951934] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] result = function(*args, **kwargs) [ 527.951934] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 527.951934] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] return func(*args, **kwargs) [ 527.951934] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 527.951934] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] raise e [ 527.951934] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.951934] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] nwinfo = self.network_api.allocate_for_instance( [ 527.951934] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.951934] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] created_port_ids = self._update_ports_for_instance( [ 527.951934] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.951934] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] with excutils.save_and_reraise_exception(): [ 527.951934] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.952320] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] self.force_reraise() [ 527.952320] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.952320] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] raise self.value [ 527.952320] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.952320] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] updated_port = self._update_port( [ 527.952320] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.952320] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] _ensure_no_port_binding_failure(port) [ 527.952320] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.952320] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] raise exception.PortBindingFailed(port_id=port['id']) [ 527.952320] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] nova.exception.PortBindingFailed: Binding failed for port 5fbe7162-22fd-463d-a62c-1725806badc1, please check neutron logs for more information. [ 527.952320] env[61906]: ERROR nova.compute.manager [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] [ 527.952634] env[61906]: DEBUG nova.compute.utils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Binding failed for port 5fbe7162-22fd-463d-a62c-1725806badc1, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 527.955179] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.457s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.959341] env[61906]: DEBUG nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Build of instance 5f417757-4295-4faa-83d2-1ad2d9c04ad6 was re-scheduled: Binding failed for port 5fbe7162-22fd-463d-a62c-1725806badc1, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 527.959938] env[61906]: DEBUG nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 527.960051] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquiring lock "refresh_cache-5f417757-4295-4faa-83d2-1ad2d9c04ad6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.960160] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquired lock "refresh_cache-5f417757-4295-4faa-83d2-1ad2d9c04ad6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.960316] env[61906]: DEBUG nova.network.neutron [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 528.302175] env[61906]: DEBUG nova.network.neutron [req-3b8ed611-b751-4fb9-a304-7b5fabcf24b5 req-f7726de2-f45c-43a5-8291-9781ccbc4389 service nova] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.350381] env[61906]: DEBUG nova.network.neutron [-] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.434753] env[61906]: DEBUG oslo_vmware.api [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Task: {'id': task-1356465, 'name': PowerOnVM_Task, 'duration_secs': 0.878406} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.435234] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 528.435355] env[61906]: DEBUG nova.compute.manager [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 528.436409] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6e4a75-8f51-4099-b12c-f26195782a73 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.527134] env[61906]: DEBUG nova.network.neutron [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.655442] env[61906]: DEBUG nova.network.neutron [req-3b8ed611-b751-4fb9-a304-7b5fabcf24b5 req-f7726de2-f45c-43a5-8291-9781ccbc4389 service nova] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.786476] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9e0f15-7d61-4026-a374-9b5a02c382b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.797987] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab27a5b-5167-4809-a0c1-93770474664c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.840273] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f677a5f-bcd1-45ad-9435-269f40414bd5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.851736] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0814420-1d14-45ae-9d50-f6dffec43925 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.856605] env[61906]: INFO nova.compute.manager [-] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Took 1.05 seconds to deallocate network for instance. [ 528.860650] env[61906]: DEBUG nova.compute.claims [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 528.860896] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.869178] env[61906]: DEBUG nova.compute.provider_tree [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.871368] env[61906]: DEBUG nova.network.neutron [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.961676] env[61906]: DEBUG oslo_concurrency.lockutils [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.160450] env[61906]: DEBUG oslo_concurrency.lockutils [req-3b8ed611-b751-4fb9-a304-7b5fabcf24b5 req-f7726de2-f45c-43a5-8291-9781ccbc4389 service nova] Releasing lock "refresh_cache-4dc3af05-b670-4f7c-bb9f-d1fa2e4180db" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.161103] env[61906]: DEBUG nova.compute.manager [req-3b8ed611-b751-4fb9-a304-7b5fabcf24b5 req-f7726de2-f45c-43a5-8291-9781ccbc4389 service nova] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Received event network-vif-deleted-7caa7da5-777c-43bf-8628-f8741f621e94 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.374354] env[61906]: DEBUG nova.scheduler.client.report [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.378031] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Releasing lock "refresh_cache-5f417757-4295-4faa-83d2-1ad2d9c04ad6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.378031] env[61906]: DEBUG nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 529.378188] env[61906]: DEBUG nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 529.378250] env[61906]: DEBUG nova.network.neutron [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 529.397988] env[61906]: DEBUG nova.network.neutron [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 529.679143] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "40f3efa6-699f-4779-b999-c86435e04c0a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.679375] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "40f3efa6-699f-4779-b999-c86435e04c0a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.881298] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.927s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.881891] env[61906]: ERROR nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port af4e9b5d-e4bf-4c0b-a45c-0e963d17f749, please check neutron logs for more information. [ 529.881891] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Traceback (most recent call last): [ 529.881891] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 529.881891] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] self.driver.spawn(context, instance, image_meta, [ 529.881891] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 529.881891] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.881891] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.881891] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] vm_ref = self.build_virtual_machine(instance, [ 529.881891] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.881891] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.881891] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.882252] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] for vif in network_info: [ 529.882252] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 529.882252] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] return self._sync_wrapper(fn, *args, **kwargs) [ 529.882252] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 529.882252] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] self.wait() [ 529.882252] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 529.882252] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] self[:] = self._gt.wait() [ 529.882252] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.882252] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] return self._exit_event.wait() [ 529.882252] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.882252] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] result = hub.switch() [ 529.882252] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.882252] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] return self.greenlet.switch() [ 529.882596] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.882596] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] result = function(*args, **kwargs) [ 529.882596] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 529.882596] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] return func(*args, **kwargs) [ 529.882596] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.882596] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] raise e [ 529.882596] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.882596] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] nwinfo = self.network_api.allocate_for_instance( [ 529.882596] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.882596] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] created_port_ids = self._update_ports_for_instance( [ 529.882596] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.882596] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] with excutils.save_and_reraise_exception(): [ 529.882596] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.882971] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] self.force_reraise() [ 529.882971] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.882971] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] raise self.value [ 529.882971] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.882971] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] updated_port = self._update_port( [ 529.882971] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.882971] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] _ensure_no_port_binding_failure(port) [ 529.882971] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.882971] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] raise exception.PortBindingFailed(port_id=port['id']) [ 529.882971] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] nova.exception.PortBindingFailed: Binding failed for port af4e9b5d-e4bf-4c0b-a45c-0e963d17f749, please check neutron logs for more information. [ 529.882971] env[61906]: ERROR nova.compute.manager [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] [ 529.883366] env[61906]: DEBUG nova.compute.utils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Binding failed for port af4e9b5d-e4bf-4c0b-a45c-0e963d17f749, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 529.883995] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.639s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.888088] env[61906]: DEBUG nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Build of instance 17b85f46-4660-4af6-825d-3f2e7ea931ca was re-scheduled: Binding failed for port af4e9b5d-e4bf-4c0b-a45c-0e963d17f749, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 529.888533] env[61906]: DEBUG nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 529.888752] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Acquiring lock "refresh_cache-17b85f46-4660-4af6-825d-3f2e7ea931ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.888895] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Acquired lock "refresh_cache-17b85f46-4660-4af6-825d-3f2e7ea931ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.890062] env[61906]: DEBUG nova.network.neutron [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 529.902245] env[61906]: DEBUG nova.network.neutron [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.406032] env[61906]: INFO nova.compute.manager [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] Took 1.03 seconds to deallocate network for instance. [ 530.421778] env[61906]: DEBUG nova.network.neutron [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 530.558996] env[61906]: DEBUG nova.network.neutron [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.581808] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquiring lock "85effb81-494c-4bda-a093-6dde523d5f45" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.582076] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Lock "85effb81-494c-4bda-a093-6dde523d5f45" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.582294] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquiring lock "85effb81-494c-4bda-a093-6dde523d5f45-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.582479] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Lock "85effb81-494c-4bda-a093-6dde523d5f45-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.582643] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Lock "85effb81-494c-4bda-a093-6dde523d5f45-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.587173] env[61906]: INFO nova.compute.manager [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Terminating instance [ 530.590346] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquiring lock "refresh_cache-85effb81-494c-4bda-a093-6dde523d5f45" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.590346] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquired lock "refresh_cache-85effb81-494c-4bda-a093-6dde523d5f45" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.591439] env[61906]: DEBUG nova.network.neutron [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 530.618962] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "081d06b3-c2b6-42f2-94b9-925493590e55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.619269] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "081d06b3-c2b6-42f2-94b9-925493590e55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.770888] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f436229e-8184-40d6-9c8f-7388456d5d3a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.782134] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2bd5be-c951-4ca1-98e2-60d93f98e4fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.826257] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c64f799-f61d-4bb7-ab34-57a8a0ccaab1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.834673] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f570688-ae70-48f6-a64f-edb695424cfd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.850681] env[61906]: DEBUG nova.compute.provider_tree [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 531.065177] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Releasing lock "refresh_cache-17b85f46-4660-4af6-825d-3f2e7ea931ca" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.065623] env[61906]: DEBUG nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 531.065811] env[61906]: DEBUG nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 531.066408] env[61906]: DEBUG nova.network.neutron [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 531.101295] env[61906]: DEBUG nova.network.neutron [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 531.128593] env[61906]: DEBUG nova.network.neutron [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 531.220872] env[61906]: DEBUG nova.network.neutron [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.250676] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Acquiring lock "ebea667f-6495-41c0-88b3-a9c61390558d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.250906] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Lock "ebea667f-6495-41c0-88b3-a9c61390558d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.375640] env[61906]: ERROR nova.scheduler.client.report [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [req-b694d2cd-d5bc-4039-a205-0261acad5612] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6e524f5d-985e-41eb-a7ae-996be3ae1c1a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b694d2cd-d5bc-4039-a205-0261acad5612"}]}: nova.exception.PortBindingFailed: Binding failed for port 61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a, please check neutron logs for more information. [ 531.395213] env[61906]: DEBUG nova.scheduler.client.report [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Refreshing inventories for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 531.418323] env[61906]: DEBUG nova.scheduler.client.report [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Updating ProviderTree inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 531.418609] env[61906]: DEBUG nova.compute.provider_tree [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 531.439359] env[61906]: DEBUG nova.scheduler.client.report [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Refreshing aggregate associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 531.452483] env[61906]: INFO nova.scheduler.client.report [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Deleted allocations for instance 5f417757-4295-4faa-83d2-1ad2d9c04ad6 [ 531.471152] env[61906]: DEBUG nova.scheduler.client.report [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Refreshing trait associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 531.603355] env[61906]: DEBUG nova.network.neutron [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.727311] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Releasing lock "refresh_cache-85effb81-494c-4bda-a093-6dde523d5f45" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.727969] env[61906]: DEBUG nova.compute.manager [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 531.728292] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 531.729718] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f81eb1f-acb3-4158-bd0c-42c3e4682a0e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.742401] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 531.742401] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-216ae57b-4aa4-4284-bf56-6d3d6d3607b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.749820] env[61906]: DEBUG oslo_vmware.api [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 531.749820] env[61906]: value = "task-1356468" [ 531.749820] env[61906]: _type = "Task" [ 531.749820] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.770833] env[61906]: DEBUG oslo_vmware.api [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356468, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.840736] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "b56dd330-d31c-4e42-baf9-d9897c55e633" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.841077] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "b56dd330-d31c-4e42-baf9-d9897c55e633" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.884266] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ccb424-6045-4e3c-8239-445f2b64f58a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.898450] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c22c50-7aa3-45af-8e4e-fbff4c602510 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.938652] env[61906]: DEBUG nova.compute.manager [None req-c5ece269-1c17-48bb-b06b-2b8e0ad86f69 tempest-ServerDiagnosticsV248Test-1288319871 tempest-ServerDiagnosticsV248Test-1288319871-project-admin] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 531.939413] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3103f09d-42e7-49c8-81a0-b388bd7aa365 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.943392] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4af146a-fbf1-4d78-a631-b4b7e2131beb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.951707] env[61906]: INFO nova.compute.manager [None req-c5ece269-1c17-48bb-b06b-2b8e0ad86f69 tempest-ServerDiagnosticsV248Test-1288319871 tempest-ServerDiagnosticsV248Test-1288319871-project-admin] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Retrieving diagnostics [ 531.954226] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce3a402-bfcf-40a2-97c6-356b9cb3828b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.957628] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7a175565-9c82-440c-b123-3b35de2e9587 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Lock "5f417757-4295-4faa-83d2-1ad2d9c04ad6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.986s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.958230] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4424306-245f-4da6-8964-56ee4973a57b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.963461] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "5f417757-4295-4faa-83d2-1ad2d9c04ad6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 19.804s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.963643] env[61906]: INFO nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 5f417757-4295-4faa-83d2-1ad2d9c04ad6] During sync_power_state the instance has a pending task (spawning). Skip. [ 531.963918] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "5f417757-4295-4faa-83d2-1ad2d9c04ad6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.999482] env[61906]: DEBUG nova.compute.provider_tree [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 532.109098] env[61906]: INFO nova.compute.manager [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] Took 1.04 seconds to deallocate network for instance. [ 532.266859] env[61906]: DEBUG oslo_vmware.api [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356468, 'name': PowerOffVM_Task, 'duration_secs': 0.125766} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.266859] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 532.267713] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 532.267713] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-856c79dd-5f70-401b-b52d-be088d14e974 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.297958] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 532.298217] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 532.298401] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Deleting the datastore file [datastore2] 85effb81-494c-4bda-a093-6dde523d5f45 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 532.301660] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2ea46e82-3e49-4c6b-9284-bd2b34e19155 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.311382] env[61906]: DEBUG oslo_vmware.api [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for the task: (returnval){ [ 532.311382] env[61906]: value = "task-1356471" [ 532.311382] env[61906]: _type = "Task" [ 532.311382] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.320354] env[61906]: DEBUG oslo_vmware.api [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356471, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.466378] env[61906]: DEBUG nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 532.557332] env[61906]: DEBUG nova.scheduler.client.report [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Updated inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with generation 33 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 532.557332] env[61906]: DEBUG nova.compute.provider_tree [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Updating resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a generation from 33 to 34 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 532.557332] env[61906]: DEBUG nova.compute.provider_tree [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 532.823736] env[61906]: DEBUG oslo_vmware.api [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Task: {'id': task-1356471, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.109101} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.826036] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 532.826036] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 532.826036] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 532.826036] env[61906]: INFO nova.compute.manager [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Took 1.10 seconds to destroy the instance on the hypervisor. [ 532.826036] env[61906]: DEBUG oslo.service.loopingcall [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 532.826036] env[61906]: DEBUG nova.compute.manager [-] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 532.826330] env[61906]: DEBUG nova.network.neutron [-] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 532.849332] env[61906]: DEBUG nova.network.neutron [-] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 533.003677] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.063132] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 3.178s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.063132] env[61906]: ERROR nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a, please check neutron logs for more information. [ 533.063132] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Traceback (most recent call last): [ 533.063132] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.063132] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] self.driver.spawn(context, instance, image_meta, [ 533.063132] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 533.063132] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.063132] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.063132] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] vm_ref = self.build_virtual_machine(instance, [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] for vif in network_info: [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] return self._sync_wrapper(fn, *args, **kwargs) [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] self.wait() [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] self[:] = self._gt.wait() [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] return self._exit_event.wait() [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.063703] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] result = hub.switch() [ 533.064427] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.064427] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] return self.greenlet.switch() [ 533.064427] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.064427] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] result = function(*args, **kwargs) [ 533.064427] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 533.064427] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] return func(*args, **kwargs) [ 533.064427] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.064427] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] raise e [ 533.064427] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.064427] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] nwinfo = self.network_api.allocate_for_instance( [ 533.064427] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.064427] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] created_port_ids = self._update_ports_for_instance( [ 533.064427] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.065884] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] with excutils.save_and_reraise_exception(): [ 533.065884] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.065884] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] self.force_reraise() [ 533.065884] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.065884] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] raise self.value [ 533.065884] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.065884] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] updated_port = self._update_port( [ 533.065884] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.065884] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] _ensure_no_port_binding_failure(port) [ 533.065884] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.065884] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] raise exception.PortBindingFailed(port_id=port['id']) [ 533.065884] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] nova.exception.PortBindingFailed: Binding failed for port 61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a, please check neutron logs for more information. [ 533.065884] env[61906]: ERROR nova.compute.manager [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] [ 533.066399] env[61906]: DEBUG nova.compute.utils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Binding failed for port 61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 533.066399] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.903s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.067170] env[61906]: INFO nova.compute.claims [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 533.073238] env[61906]: DEBUG nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Build of instance 2f64e954-bf3e-464f-a72e-d83f0783af5c was re-scheduled: Binding failed for port 61a70238-9fa2-4d0d-b1a9-c7ccb3c7769a, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 533.074669] env[61906]: DEBUG nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 533.074669] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "refresh_cache-2f64e954-bf3e-464f-a72e-d83f0783af5c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.078313] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "refresh_cache-2f64e954-bf3e-464f-a72e-d83f0783af5c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.079197] env[61906]: DEBUG nova.network.neutron [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 533.187186] env[61906]: INFO nova.scheduler.client.report [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Deleted allocations for instance 17b85f46-4660-4af6-825d-3f2e7ea931ca [ 533.351679] env[61906]: DEBUG nova.network.neutron [-] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.505801] env[61906]: DEBUG oslo_concurrency.lockutils [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Acquiring lock "7f4bb9fb-fa64-451d-9876-7e4314c13a64" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.506911] env[61906]: DEBUG oslo_concurrency.lockutils [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Lock "7f4bb9fb-fa64-451d-9876-7e4314c13a64" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.507191] env[61906]: DEBUG oslo_concurrency.lockutils [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Acquiring lock "7f4bb9fb-fa64-451d-9876-7e4314c13a64-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.507407] env[61906]: DEBUG oslo_concurrency.lockutils [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Lock "7f4bb9fb-fa64-451d-9876-7e4314c13a64-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.507577] env[61906]: DEBUG oslo_concurrency.lockutils [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Lock "7f4bb9fb-fa64-451d-9876-7e4314c13a64-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.509826] env[61906]: INFO nova.compute.manager [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Terminating instance [ 533.513881] env[61906]: DEBUG oslo_concurrency.lockutils [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Acquiring lock "refresh_cache-7f4bb9fb-fa64-451d-9876-7e4314c13a64" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.514045] env[61906]: DEBUG oslo_concurrency.lockutils [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Acquired lock "refresh_cache-7f4bb9fb-fa64-451d-9876-7e4314c13a64" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.514210] env[61906]: DEBUG nova.network.neutron [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 533.623799] env[61906]: DEBUG nova.network.neutron [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 533.698565] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4a684a03-c267-45aa-838b-52034e4ace3a tempest-ServersTestJSON-500715912 tempest-ServersTestJSON-500715912-project-member] Lock "17b85f46-4660-4af6-825d-3f2e7ea931ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.059s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.700117] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "17b85f46-4660-4af6-825d-3f2e7ea931ca" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 21.540s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.700315] env[61906]: INFO nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 17b85f46-4660-4af6-825d-3f2e7ea931ca] During sync_power_state the instance has a pending task (spawning). Skip. [ 533.700486] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "17b85f46-4660-4af6-825d-3f2e7ea931ca" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.856406] env[61906]: INFO nova.compute.manager [-] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Took 1.03 seconds to deallocate network for instance. [ 534.070622] env[61906]: DEBUG nova.network.neutron [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.078030] env[61906]: DEBUG nova.network.neutron [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.208156] env[61906]: DEBUG nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 534.311931] env[61906]: DEBUG nova.network.neutron [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.366602] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.527968] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818b6acf-f7ee-45ed-90b9-aea8849dc34c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.544863] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f151d734-6c4c-4e58-8667-414c62d9b053 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.586523] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "refresh_cache-2f64e954-bf3e-464f-a72e-d83f0783af5c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.586882] env[61906]: DEBUG nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 534.586950] env[61906]: DEBUG nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 534.587125] env[61906]: DEBUG nova.network.neutron [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 534.589941] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8bd30d-897b-4165-a106-d508150b6664 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.602522] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48082ecd-f4ec-4159-9861-720aae1047e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.619379] env[61906]: DEBUG nova.compute.provider_tree [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.667590] env[61906]: DEBUG nova.network.neutron [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.736053] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.815997] env[61906]: DEBUG oslo_concurrency.lockutils [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Releasing lock "refresh_cache-7f4bb9fb-fa64-451d-9876-7e4314c13a64" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.819109] env[61906]: DEBUG nova.compute.manager [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 534.819109] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 534.819109] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b47871d-052f-4379-b20d-ea412dc9c6d9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.832319] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 534.832603] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-51f5c212-aa36-44c6-a353-69abb9f4acba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.843652] env[61906]: DEBUG oslo_vmware.api [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Waiting for the task: (returnval){ [ 534.843652] env[61906]: value = "task-1356473" [ 534.843652] env[61906]: _type = "Task" [ 534.843652] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.857598] env[61906]: DEBUG oslo_vmware.api [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356473, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.122805] env[61906]: DEBUG nova.scheduler.client.report [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 535.171455] env[61906]: DEBUG nova.network.neutron [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.241484] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquiring lock "f9edb738-210e-4f85-8062-7759fa9033ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.241484] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Lock "f9edb738-210e-4f85-8062-7759fa9033ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.354183] env[61906]: DEBUG oslo_vmware.api [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356473, 'name': PowerOffVM_Task, 'duration_secs': 0.174218} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.354524] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 535.354654] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 535.354868] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f86cf0a-cc93-4c9a-856a-43d31e566e63 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.385304] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 535.387677] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 535.388943] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Deleting the datastore file [datastore2] 7f4bb9fb-fa64-451d-9876-7e4314c13a64 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 535.388943] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-23cb6735-7811-417f-92ac-3c68b0f50285 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.397930] env[61906]: DEBUG oslo_vmware.api [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Waiting for the task: (returnval){ [ 535.397930] env[61906]: value = "task-1356475" [ 535.397930] env[61906]: _type = "Task" [ 535.397930] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.411613] env[61906]: DEBUG oslo_vmware.api [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356475, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.633927] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.634396] env[61906]: DEBUG nova.compute.manager [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 535.638264] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.950s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.640080] env[61906]: INFO nova.compute.claims [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 535.675503] env[61906]: INFO nova.compute.manager [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] Took 1.09 seconds to deallocate network for instance. [ 535.911334] env[61906]: DEBUG oslo_vmware.api [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Task: {'id': task-1356475, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210447} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.912537] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 535.912537] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 535.912537] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 535.912537] env[61906]: INFO nova.compute.manager [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Took 1.10 seconds to destroy the instance on the hypervisor. [ 535.912537] env[61906]: DEBUG oslo.service.loopingcall [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.912791] env[61906]: DEBUG nova.compute.manager [-] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.912791] env[61906]: DEBUG nova.network.neutron [-] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 535.948547] env[61906]: DEBUG nova.network.neutron [-] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.144964] env[61906]: DEBUG nova.compute.utils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 536.151177] env[61906]: DEBUG nova.compute.manager [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 536.451875] env[61906]: DEBUG nova.network.neutron [-] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.608445] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Acquiring lock "3e6732aa-a878-4bde-9299-abd779f2a109" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.608445] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Lock "3e6732aa-a878-4bde-9299-abd779f2a109" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.652218] env[61906]: DEBUG nova.compute.manager [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 536.732020] env[61906]: INFO nova.scheduler.client.report [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Deleted allocations for instance 2f64e954-bf3e-464f-a72e-d83f0783af5c [ 536.956427] env[61906]: INFO nova.compute.manager [-] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Took 1.04 seconds to deallocate network for instance. [ 537.067274] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c690a478-52c3-4f68-9e57-a85d799a3670 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.076890] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38928da4-b38c-4953-9ec4-2278de4fa3a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.113141] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9f89bad-43c0-4251-807e-6aa502e7529a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.123256] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c2bdbc-bd1a-4525-959b-054bd17feed1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.144809] env[61906]: DEBUG nova.compute.provider_tree [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 537.242924] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9ab8b830-ca79-4879-b813-725035b68a7c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "2f64e954-bf3e-464f-a72e-d83f0783af5c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.540s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.243682] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "2f64e954-bf3e-464f-a72e-d83f0783af5c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 25.084s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.243881] env[61906]: INFO nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 2f64e954-bf3e-464f-a72e-d83f0783af5c] During sync_power_state the instance has a pending task (spawning). Skip. [ 537.244199] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "2f64e954-bf3e-464f-a72e-d83f0783af5c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.466764] env[61906]: DEBUG oslo_concurrency.lockutils [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.647128] env[61906]: DEBUG nova.scheduler.client.report [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 537.666241] env[61906]: DEBUG nova.compute.manager [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 537.696100] env[61906]: DEBUG nova.virt.hardware [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 537.696406] env[61906]: DEBUG nova.virt.hardware [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 537.696602] env[61906]: DEBUG nova.virt.hardware [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 537.696710] env[61906]: DEBUG nova.virt.hardware [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 537.696797] env[61906]: DEBUG nova.virt.hardware [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 537.697164] env[61906]: DEBUG nova.virt.hardware [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 537.699650] env[61906]: DEBUG nova.virt.hardware [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 537.699650] env[61906]: DEBUG nova.virt.hardware [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 537.699650] env[61906]: DEBUG nova.virt.hardware [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 537.699650] env[61906]: DEBUG nova.virt.hardware [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 537.699650] env[61906]: DEBUG nova.virt.hardware [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 537.700380] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe880a97-aab5-4f84-8331-bacb17e2d5dc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.712952] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6835c52b-1c34-4473-a3fc-f8d335acd946 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.729967] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 537.740218] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Creating folder: Project (f1f01f2f82ae44918605f3018e65b512). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 537.740218] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3deaf1af-7e13-4f61-808f-a73fb8452c92 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.747020] env[61906]: DEBUG nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 537.756224] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Created folder: Project (f1f01f2f82ae44918605f3018e65b512) in parent group-v288914. [ 537.756473] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Creating folder: Instances. Parent ref: group-v288927. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 537.757041] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e481595e-0d43-4159-b58e-70a80c8de574 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.770367] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Created folder: Instances in parent group-v288927. [ 537.770367] env[61906]: DEBUG oslo.service.loopingcall [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 537.770367] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 537.770367] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0613a8eb-ec6e-40c7-9265-aacc811cd23f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.796246] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 537.796246] env[61906]: value = "task-1356478" [ 537.796246] env[61906]: _type = "Task" [ 537.796246] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.807660] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356478, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.152053] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.152898] env[61906]: DEBUG nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 538.155347] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.972s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.156850] env[61906]: INFO nova.compute.claims [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 538.291333] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.315135] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356478, 'name': CreateVM_Task, 'duration_secs': 0.311296} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.315135] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 538.315582] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.317129] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.317129] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 538.317129] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d747d56-6ca3-47e3-aea2-759b6dfbc716 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.322123] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Waiting for the task: (returnval){ [ 538.322123] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52008d82-9021-f3bb-3ca9-1e8cdb24b7b2" [ 538.322123] env[61906]: _type = "Task" [ 538.322123] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.331693] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52008d82-9021-f3bb-3ca9-1e8cdb24b7b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.670479] env[61906]: DEBUG nova.compute.utils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 538.675132] env[61906]: DEBUG nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 538.675132] env[61906]: DEBUG nova.network.neutron [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 538.792554] env[61906]: DEBUG nova.policy [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b546331e0d1f44cbb9b55967b222d316', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f376d1ed4095454cb7e17227af1449bb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 538.862013] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52008d82-9021-f3bb-3ca9-1e8cdb24b7b2, 'name': SearchDatastore_Task, 'duration_secs': 0.014785} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.862013] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.862013] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 538.862013] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.862346] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.862346] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 538.862799] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-854dcaec-c446-4436-95ad-03165df973b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.880257] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 538.880554] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 538.882802] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63944757-9167-4540-9be2-d52482712a86 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.896131] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Waiting for the task: (returnval){ [ 538.896131] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fd8c2c-379a-d979-48a5-d3813cd43b7f" [ 538.896131] env[61906]: _type = "Task" [ 538.896131] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.918893] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fd8c2c-379a-d979-48a5-d3813cd43b7f, 'name': SearchDatastore_Task, 'duration_secs': 0.010744} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.920576] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1139499-fe47-44cf-be04-77b9228904ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.931189] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Waiting for the task: (returnval){ [ 538.931189] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527c2ead-db5f-b5db-782f-9543eec46426" [ 538.931189] env[61906]: _type = "Task" [ 538.931189] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.949201] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527c2ead-db5f-b5db-782f-9543eec46426, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.179146] env[61906]: DEBUG nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 539.379494] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Acquiring lock "73a96c0e-a08b-46c9-b700-25544e17c73d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.379611] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Lock "73a96c0e-a08b-46c9-b700-25544e17c73d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.440408] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527c2ead-db5f-b5db-782f-9543eec46426, 'name': SearchDatastore_Task, 'duration_secs': 0.01134} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.440823] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.440934] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] e23fd777-66f7-4ee1-bdcb-fe19c55f955c/e23fd777-66f7-4ee1-bdcb-fe19c55f955c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 539.441117] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8a843a9d-caaa-4ce8-ac73-e0ac46c5ffa9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.451190] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Waiting for the task: (returnval){ [ 539.451190] env[61906]: value = "task-1356479" [ 539.451190] env[61906]: _type = "Task" [ 539.451190] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.461642] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356479, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.570217] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "0c7d1af9-54e9-435d-a0cc-687f793d4f43" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.570456] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "0c7d1af9-54e9-435d-a0cc-687f793d4f43" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.606389] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "7aee8304-8d07-4aa2-867a-2eeeeacbabc8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.606389] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "7aee8304-8d07-4aa2-867a-2eeeeacbabc8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.631107] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b04be2e-40eb-457b-99be-ddbb24ab7b86 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.639592] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.642051] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.644999] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c155d08c-d714-403d-a416-2c8b9a2b664b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.680223] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d551a14-01c4-496d-a297-a202ce5c7aa3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.697269] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6161c9-6eaf-4426-bb43-a9e27c1faf8d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.713429] env[61906]: DEBUG nova.compute.provider_tree [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.963225] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356479, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.081241] env[61906]: DEBUG nova.network.neutron [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Successfully created port: 15d04bfa-35cb-4734-9244-a4ecfe472fa3 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 540.148815] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "569d7ab1-22b8-4006-bf0e-ff35032cb9a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.148815] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "569d7ab1-22b8-4006-bf0e-ff35032cb9a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.199079] env[61906]: DEBUG nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 540.218672] env[61906]: DEBUG nova.scheduler.client.report [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 540.235039] env[61906]: DEBUG nova.virt.hardware [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 540.235039] env[61906]: DEBUG nova.virt.hardware [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 540.235039] env[61906]: DEBUG nova.virt.hardware [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 540.235521] env[61906]: DEBUG nova.virt.hardware [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 540.235521] env[61906]: DEBUG nova.virt.hardware [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 540.235521] env[61906]: DEBUG nova.virt.hardware [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 540.235853] env[61906]: DEBUG nova.virt.hardware [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 540.235912] env[61906]: DEBUG nova.virt.hardware [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 540.237267] env[61906]: DEBUG nova.virt.hardware [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 540.237267] env[61906]: DEBUG nova.virt.hardware [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 540.237267] env[61906]: DEBUG nova.virt.hardware [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 540.239128] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9ecaf7-8c2f-41f6-89da-21a8f61901e2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.251050] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d556b98e-fcac-4a14-8fd1-fa6b76f79bc3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.461741] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356479, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518899} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.462024] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] e23fd777-66f7-4ee1-bdcb-fe19c55f955c/e23fd777-66f7-4ee1-bdcb-fe19c55f955c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 540.464353] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 540.464353] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-08300574-0bf7-464a-ae28-4325ce8a287e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.472023] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Waiting for the task: (returnval){ [ 540.472023] env[61906]: value = "task-1356480" [ 540.472023] env[61906]: _type = "Task" [ 540.472023] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.483646] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356480, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.730858] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.730858] env[61906]: DEBUG nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 540.733261] env[61906]: DEBUG oslo_concurrency.lockutils [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.370s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.733611] env[61906]: DEBUG nova.objects.instance [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61906) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 540.986018] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356480, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064513} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.986018] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 540.986018] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265dc0ad-4e7a-4be4-91d7-e3a2bd9ee234 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.015853] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Reconfiguring VM instance instance-0000000c to attach disk [datastore2] e23fd777-66f7-4ee1-bdcb-fe19c55f955c/e23fd777-66f7-4ee1-bdcb-fe19c55f955c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 541.015853] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2ecbbcc-9e35-43e4-962b-37266bb3346f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.043173] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Waiting for the task: (returnval){ [ 541.043173] env[61906]: value = "task-1356481" [ 541.043173] env[61906]: _type = "Task" [ 541.043173] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.052686] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356481, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.206788] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquiring lock "e5c81f8b-aac3-4cf8-9abb-444731f57206" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.207546] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Lock "e5c81f8b-aac3-4cf8-9abb-444731f57206" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.238396] env[61906]: DEBUG nova.compute.utils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 541.252198] env[61906]: DEBUG nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 541.252198] env[61906]: DEBUG nova.network.neutron [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 541.556966] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356481, 'name': ReconfigVM_Task, 'duration_secs': 0.293147} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.557268] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Reconfigured VM instance instance-0000000c to attach disk [datastore2] e23fd777-66f7-4ee1-bdcb-fe19c55f955c/e23fd777-66f7-4ee1-bdcb-fe19c55f955c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 541.559350] env[61906]: DEBUG nova.policy [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'abd858787f6945349a08939438d73518', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fde398d0f99d4033b15e86d479855c50', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 541.562184] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-797b90ce-e540-4252-bdb3-3006959c206e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.571118] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Waiting for the task: (returnval){ [ 541.571118] env[61906]: value = "task-1356482" [ 541.571118] env[61906]: _type = "Task" [ 541.571118] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.582765] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356482, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.759246] env[61906]: DEBUG nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 541.760585] env[61906]: DEBUG oslo_concurrency.lockutils [None req-365d1d06-b5ba-4a76-9b3f-ddc53bc74ae2 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.027s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.766026] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 21.118s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.083887] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356482, 'name': Rename_Task, 'duration_secs': 0.167762} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.084640] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 542.084911] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10160fae-44ad-4c1c-a129-ac2184148ce7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.092995] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Waiting for the task: (returnval){ [ 542.092995] env[61906]: value = "task-1356483" [ 542.092995] env[61906]: _type = "Task" [ 542.092995] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.106922] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356483, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.326765] env[61906]: DEBUG nova.network.neutron [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Successfully created port: 3cee9f61-8537-4ca1-bef4-683dfe02020c {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 542.611461] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356483, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.787913] env[61906]: DEBUG nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 542.804193] env[61906]: ERROR nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 15d04bfa-35cb-4734-9244-a4ecfe472fa3, please check neutron logs for more information. [ 542.804193] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 542.804193] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.804193] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 542.804193] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 542.804193] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 542.804193] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 542.804193] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 542.804193] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.804193] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 542.804193] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.804193] env[61906]: ERROR nova.compute.manager raise self.value [ 542.804193] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 542.804193] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 542.804193] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.804193] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 542.804774] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.804774] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 542.804774] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 15d04bfa-35cb-4734-9244-a4ecfe472fa3, please check neutron logs for more information. [ 542.804774] env[61906]: ERROR nova.compute.manager [ 542.804774] env[61906]: Traceback (most recent call last): [ 542.804774] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 542.804774] env[61906]: listener.cb(fileno) [ 542.804774] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.804774] env[61906]: result = function(*args, **kwargs) [ 542.804774] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 542.804774] env[61906]: return func(*args, **kwargs) [ 542.804774] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.804774] env[61906]: raise e [ 542.804774] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.804774] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 542.804774] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 542.804774] env[61906]: created_port_ids = self._update_ports_for_instance( [ 542.804774] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 542.804774] env[61906]: with excutils.save_and_reraise_exception(): [ 542.804774] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.804774] env[61906]: self.force_reraise() [ 542.804774] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.804774] env[61906]: raise self.value [ 542.804774] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 542.804774] env[61906]: updated_port = self._update_port( [ 542.804774] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.804774] env[61906]: _ensure_no_port_binding_failure(port) [ 542.804774] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.804774] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 542.806934] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 15d04bfa-35cb-4734-9244-a4ecfe472fa3, please check neutron logs for more information. [ 542.806934] env[61906]: Removing descriptor: 16 [ 542.806934] env[61906]: ERROR nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 15d04bfa-35cb-4734-9244-a4ecfe472fa3, please check neutron logs for more information. [ 542.806934] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Traceback (most recent call last): [ 542.806934] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 542.806934] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] yield resources [ 542.806934] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 542.806934] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] self.driver.spawn(context, instance, image_meta, [ 542.806934] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 542.806934] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 542.806934] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 542.806934] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] vm_ref = self.build_virtual_machine(instance, [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] for vif in network_info: [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] return self._sync_wrapper(fn, *args, **kwargs) [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] self.wait() [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] self[:] = self._gt.wait() [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] return self._exit_event.wait() [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 542.807656] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] result = hub.switch() [ 542.808131] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 542.808131] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] return self.greenlet.switch() [ 542.808131] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.808131] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] result = function(*args, **kwargs) [ 542.808131] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 542.808131] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] return func(*args, **kwargs) [ 542.808131] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.808131] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] raise e [ 542.808131] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.808131] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] nwinfo = self.network_api.allocate_for_instance( [ 542.808131] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 542.808131] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] created_port_ids = self._update_ports_for_instance( [ 542.808131] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 542.808822] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] with excutils.save_and_reraise_exception(): [ 542.808822] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.808822] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] self.force_reraise() [ 542.808822] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.808822] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] raise self.value [ 542.808822] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 542.808822] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] updated_port = self._update_port( [ 542.808822] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.808822] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] _ensure_no_port_binding_failure(port) [ 542.808822] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.808822] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] raise exception.PortBindingFailed(port_id=port['id']) [ 542.808822] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] nova.exception.PortBindingFailed: Binding failed for port 15d04bfa-35cb-4734-9244-a4ecfe472fa3, please check neutron logs for more information. [ 542.808822] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] [ 542.809230] env[61906]: INFO nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Terminating instance [ 542.810718] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Acquiring lock "refresh_cache-6c354d85-0252-4eb9-81e2-550a403bdbc5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.811340] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Acquired lock "refresh_cache-6c354d85-0252-4eb9-81e2-550a403bdbc5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.811340] env[61906]: DEBUG nova.network.neutron [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 542.826624] env[61906]: WARNING nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 85effb81-494c-4bda-a093-6dde523d5f45 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 542.827152] env[61906]: WARNING nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 7f4bb9fb-fa64-451d-9876-7e4314c13a64 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 542.827288] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance fa0cbe6c-937b-4276-a419-0ab6e92e795f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.827407] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 9f6864a2-3e54-41cb-b858-16e6e350a2af actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.827522] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.827633] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance e23fd777-66f7-4ee1-bdcb-fe19c55f955c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.827742] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 6c354d85-0252-4eb9-81e2-550a403bdbc5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.829143] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance f708ec4a-7f1e-4bfa-b561-4b9bc5556d27 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 542.844226] env[61906]: DEBUG nova.virt.hardware [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 542.844512] env[61906]: DEBUG nova.virt.hardware [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 542.844688] env[61906]: DEBUG nova.virt.hardware [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 542.844873] env[61906]: DEBUG nova.virt.hardware [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 542.845565] env[61906]: DEBUG nova.virt.hardware [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 542.845802] env[61906]: DEBUG nova.virt.hardware [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 542.846034] env[61906]: DEBUG nova.virt.hardware [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 542.846221] env[61906]: DEBUG nova.virt.hardware [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 542.846534] env[61906]: DEBUG nova.virt.hardware [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 542.846572] env[61906]: DEBUG nova.virt.hardware [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 542.846715] env[61906]: DEBUG nova.virt.hardware [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 542.848980] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c7f03b-42cc-4671-a1d5-556b787d5f71 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.863580] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1760ec72-aea9-4daa-948d-f8187ae8a656 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.070056] env[61906]: DEBUG nova.compute.manager [req-62908c34-bfe6-4ecd-9b74-a76b914555c8 req-e4522092-7673-4d63-ae62-1ef33c19de4b service nova] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Received event network-changed-15d04bfa-35cb-4734-9244-a4ecfe472fa3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 543.070311] env[61906]: DEBUG nova.compute.manager [req-62908c34-bfe6-4ecd-9b74-a76b914555c8 req-e4522092-7673-4d63-ae62-1ef33c19de4b service nova] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Refreshing instance network info cache due to event network-changed-15d04bfa-35cb-4734-9244-a4ecfe472fa3. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 543.070547] env[61906]: DEBUG oslo_concurrency.lockutils [req-62908c34-bfe6-4ecd-9b74-a76b914555c8 req-e4522092-7673-4d63-ae62-1ef33c19de4b service nova] Acquiring lock "refresh_cache-6c354d85-0252-4eb9-81e2-550a403bdbc5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.098274] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Acquiring lock "95604a75-0ab3-45d9-a235-9067e615e72a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.098613] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Lock "95604a75-0ab3-45d9-a235-9067e615e72a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.115646] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356483, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.336849] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance ec81c4ec-07c6-4b1f-8555-69902cff4765 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 543.347982] env[61906]: DEBUG nova.network.neutron [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.512053] env[61906]: DEBUG nova.network.neutron [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.614059] env[61906]: DEBUG oslo_vmware.api [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356483, 'name': PowerOnVM_Task, 'duration_secs': 1.268207} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.614059] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 543.614059] env[61906]: INFO nova.compute.manager [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Took 5.95 seconds to spawn the instance on the hypervisor. [ 543.614059] env[61906]: DEBUG nova.compute.manager [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 543.614059] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c83e7ba-edf4-4c7f-8279-e2b82796c712 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.845178] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance f72e6538-7130-45a9-a722-ba60de3bda49 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 544.015514] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Releasing lock "refresh_cache-6c354d85-0252-4eb9-81e2-550a403bdbc5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.016061] env[61906]: DEBUG nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 544.016324] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 544.016662] env[61906]: DEBUG oslo_concurrency.lockutils [req-62908c34-bfe6-4ecd-9b74-a76b914555c8 req-e4522092-7673-4d63-ae62-1ef33c19de4b service nova] Acquired lock "refresh_cache-6c354d85-0252-4eb9-81e2-550a403bdbc5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.016848] env[61906]: DEBUG nova.network.neutron [req-62908c34-bfe6-4ecd-9b74-a76b914555c8 req-e4522092-7673-4d63-ae62-1ef33c19de4b service nova] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Refreshing network info cache for port 15d04bfa-35cb-4734-9244-a4ecfe472fa3 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 544.018171] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4c5976ff-c471-40e8-acf2-4c88e3b41d2e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.029879] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aea66bb-5bdf-4330-bfcb-97500a035fc4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.061279] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6c354d85-0252-4eb9-81e2-550a403bdbc5 could not be found. [ 544.064847] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 544.064847] env[61906]: INFO nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Took 0.05 seconds to destroy the instance on the hypervisor. [ 544.064847] env[61906]: DEBUG oslo.service.loopingcall [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 544.064847] env[61906]: DEBUG nova.compute.manager [-] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 544.064847] env[61906]: DEBUG nova.network.neutron [-] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 544.105235] env[61906]: DEBUG nova.network.neutron [-] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.134020] env[61906]: INFO nova.compute.manager [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Took 27.00 seconds to build instance. [ 544.347616] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 6b3ecea4-fe20-4b6c-a987-59c20731b9ac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 544.407335] env[61906]: ERROR nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3cee9f61-8537-4ca1-bef4-683dfe02020c, please check neutron logs for more information. [ 544.407335] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 544.407335] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.407335] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 544.407335] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 544.407335] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 544.407335] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 544.407335] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 544.407335] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.407335] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 544.407335] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.407335] env[61906]: ERROR nova.compute.manager raise self.value [ 544.407335] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 544.407335] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 544.407335] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.407335] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 544.408779] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.408779] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 544.408779] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3cee9f61-8537-4ca1-bef4-683dfe02020c, please check neutron logs for more information. [ 544.408779] env[61906]: ERROR nova.compute.manager [ 544.408779] env[61906]: Traceback (most recent call last): [ 544.408779] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 544.408779] env[61906]: listener.cb(fileno) [ 544.408779] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.408779] env[61906]: result = function(*args, **kwargs) [ 544.408779] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 544.408779] env[61906]: return func(*args, **kwargs) [ 544.408779] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.408779] env[61906]: raise e [ 544.408779] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.408779] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 544.408779] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 544.408779] env[61906]: created_port_ids = self._update_ports_for_instance( [ 544.408779] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 544.408779] env[61906]: with excutils.save_and_reraise_exception(): [ 544.408779] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.408779] env[61906]: self.force_reraise() [ 544.408779] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.408779] env[61906]: raise self.value [ 544.408779] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 544.408779] env[61906]: updated_port = self._update_port( [ 544.408779] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.408779] env[61906]: _ensure_no_port_binding_failure(port) [ 544.408779] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.408779] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 544.409608] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 3cee9f61-8537-4ca1-bef4-683dfe02020c, please check neutron logs for more information. [ 544.409608] env[61906]: Removing descriptor: 17 [ 544.409608] env[61906]: ERROR nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3cee9f61-8537-4ca1-bef4-683dfe02020c, please check neutron logs for more information. [ 544.409608] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Traceback (most recent call last): [ 544.409608] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 544.409608] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] yield resources [ 544.409608] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.409608] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] self.driver.spawn(context, instance, image_meta, [ 544.409608] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 544.409608] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.409608] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.409608] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] vm_ref = self.build_virtual_machine(instance, [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] for vif in network_info: [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] return self._sync_wrapper(fn, *args, **kwargs) [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] self.wait() [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] self[:] = self._gt.wait() [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] return self._exit_event.wait() [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.410049] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] result = hub.switch() [ 544.410509] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.410509] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] return self.greenlet.switch() [ 544.410509] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.410509] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] result = function(*args, **kwargs) [ 544.410509] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 544.410509] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] return func(*args, **kwargs) [ 544.410509] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.410509] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] raise e [ 544.410509] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.410509] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] nwinfo = self.network_api.allocate_for_instance( [ 544.410509] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 544.410509] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] created_port_ids = self._update_ports_for_instance( [ 544.410509] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 544.410944] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] with excutils.save_and_reraise_exception(): [ 544.410944] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.410944] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] self.force_reraise() [ 544.410944] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.410944] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] raise self.value [ 544.410944] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 544.410944] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] updated_port = self._update_port( [ 544.410944] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.410944] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] _ensure_no_port_binding_failure(port) [ 544.410944] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.410944] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] raise exception.PortBindingFailed(port_id=port['id']) [ 544.410944] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] nova.exception.PortBindingFailed: Binding failed for port 3cee9f61-8537-4ca1-bef4-683dfe02020c, please check neutron logs for more information. [ 544.410944] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] [ 544.411514] env[61906]: INFO nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Terminating instance [ 544.411514] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquiring lock "refresh_cache-f708ec4a-7f1e-4bfa-b561-4b9bc5556d27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.411514] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquired lock "refresh_cache-f708ec4a-7f1e-4bfa-b561-4b9bc5556d27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.411514] env[61906]: DEBUG nova.network.neutron [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 544.543212] env[61906]: DEBUG nova.network.neutron [req-62908c34-bfe6-4ecd-9b74-a76b914555c8 req-e4522092-7673-4d63-ae62-1ef33c19de4b service nova] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.609862] env[61906]: DEBUG nova.network.neutron [-] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.635166] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9e28cda8-717c-4211-88a2-84a6cc97e2aa tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Lock "e23fd777-66f7-4ee1-bdcb-fe19c55f955c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.168s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.693970] env[61906]: DEBUG nova.network.neutron [req-62908c34-bfe6-4ecd-9b74-a76b914555c8 req-e4522092-7673-4d63-ae62-1ef33c19de4b service nova] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.855635] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance d23d0d19-4806-4d5e-aa05-13aaf3c4a770 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 545.006405] env[61906]: DEBUG nova.network.neutron [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.066666] env[61906]: DEBUG nova.compute.manager [None req-92c34b61-f45d-4399-9e5e-4e6545f12f75 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 545.067219] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d032fa00-91cd-43bb-a697-4ef69df2c534 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.113199] env[61906]: INFO nova.compute.manager [-] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Took 1.05 seconds to deallocate network for instance. [ 545.120718] env[61906]: DEBUG nova.compute.claims [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 545.120930] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.133027] env[61906]: DEBUG nova.network.neutron [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.140668] env[61906]: DEBUG nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 545.198091] env[61906]: DEBUG oslo_concurrency.lockutils [req-62908c34-bfe6-4ecd-9b74-a76b914555c8 req-e4522092-7673-4d63-ae62-1ef33c19de4b service nova] Releasing lock "refresh_cache-6c354d85-0252-4eb9-81e2-550a403bdbc5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.237210] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "42135fac-aef7-47d9-9afe-5ac2351bb07c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.237210] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "42135fac-aef7-47d9-9afe-5ac2351bb07c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.359421] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 528a2095-e82e-44eb-a4a1-dacde88966bb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 545.591650] env[61906]: INFO nova.compute.manager [None req-92c34b61-f45d-4399-9e5e-4e6545f12f75 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] instance snapshotting [ 545.592311] env[61906]: DEBUG nova.objects.instance [None req-92c34b61-f45d-4399-9e5e-4e6545f12f75 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Lazy-loading 'flavor' on Instance uuid e23fd777-66f7-4ee1-bdcb-fe19c55f955c {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 545.636930] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Releasing lock "refresh_cache-f708ec4a-7f1e-4bfa-b561-4b9bc5556d27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.636930] env[61906]: DEBUG nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 545.636930] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 545.637729] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c043f1ea-0757-48e2-bfab-9653267569fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.650926] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea9148b-0b2e-4b7d-93d5-a3a5de490d64 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.670935] env[61906]: DEBUG nova.compute.manager [req-5b3f4390-d58c-46ba-a943-14e838260ec5 req-3224b7ea-d9a7-41a4-abc5-cb58adee6174 service nova] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Received event network-vif-deleted-15d04bfa-35cb-4734-9244-a4ecfe472fa3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 545.677303] env[61906]: DEBUG nova.compute.manager [req-5b3f4390-d58c-46ba-a943-14e838260ec5 req-3224b7ea-d9a7-41a4-abc5-cb58adee6174 service nova] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Received event network-changed-3cee9f61-8537-4ca1-bef4-683dfe02020c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 545.677303] env[61906]: DEBUG nova.compute.manager [req-5b3f4390-d58c-46ba-a943-14e838260ec5 req-3224b7ea-d9a7-41a4-abc5-cb58adee6174 service nova] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Refreshing instance network info cache due to event network-changed-3cee9f61-8537-4ca1-bef4-683dfe02020c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 545.677303] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b3f4390-d58c-46ba-a943-14e838260ec5 req-3224b7ea-d9a7-41a4-abc5-cb58adee6174 service nova] Acquiring lock "refresh_cache-f708ec4a-7f1e-4bfa-b561-4b9bc5556d27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.677303] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b3f4390-d58c-46ba-a943-14e838260ec5 req-3224b7ea-d9a7-41a4-abc5-cb58adee6174 service nova] Acquired lock "refresh_cache-f708ec4a-7f1e-4bfa-b561-4b9bc5556d27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.677303] env[61906]: DEBUG nova.network.neutron [req-5b3f4390-d58c-46ba-a943-14e838260ec5 req-3224b7ea-d9a7-41a4-abc5-cb58adee6174 service nova] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Refreshing network info cache for port 3cee9f61-8537-4ca1-bef4-683dfe02020c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 545.687425] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.697437] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f708ec4a-7f1e-4bfa-b561-4b9bc5556d27 could not be found. [ 545.697658] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 545.697959] env[61906]: INFO nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Took 0.06 seconds to destroy the instance on the hypervisor. [ 545.698150] env[61906]: DEBUG oslo.service.loopingcall [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 545.702040] env[61906]: DEBUG nova.compute.manager [-] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.702040] env[61906]: DEBUG nova.network.neutron [-] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 545.736105] env[61906]: DEBUG nova.network.neutron [-] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.865844] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance e64fd277-10ae-494d-918a-2f3faa4e44d1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 545.867925] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Acquiring lock "e23fd777-66f7-4ee1-bdcb-fe19c55f955c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.868174] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Lock "e23fd777-66f7-4ee1-bdcb-fe19c55f955c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.868375] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Acquiring lock "e23fd777-66f7-4ee1-bdcb-fe19c55f955c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.868555] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Lock "e23fd777-66f7-4ee1-bdcb-fe19c55f955c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.868823] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Lock "e23fd777-66f7-4ee1-bdcb-fe19c55f955c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.870658] env[61906]: INFO nova.compute.manager [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Terminating instance [ 545.873117] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Acquiring lock "refresh_cache-e23fd777-66f7-4ee1-bdcb-fe19c55f955c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.873235] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Acquired lock "refresh_cache-e23fd777-66f7-4ee1-bdcb-fe19c55f955c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.873589] env[61906]: DEBUG nova.network.neutron [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 546.101018] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-388e7a96-42d1-40f8-8cc3-c570e7f959b1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.123158] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9707cad1-a24b-47aa-8535-d04182daab5d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.202726] env[61906]: DEBUG nova.network.neutron [req-5b3f4390-d58c-46ba-a943-14e838260ec5 req-3224b7ea-d9a7-41a4-abc5-cb58adee6174 service nova] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.242663] env[61906]: DEBUG nova.network.neutron [-] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.324543] env[61906]: DEBUG nova.network.neutron [req-5b3f4390-d58c-46ba-a943-14e838260ec5 req-3224b7ea-d9a7-41a4-abc5-cb58adee6174 service nova] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.370021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 4b0abc14-0ce2-4212-8677-82f6279ee107 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 546.409404] env[61906]: DEBUG nova.network.neutron [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.475361] env[61906]: DEBUG nova.network.neutron [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.636227] env[61906]: DEBUG nova.compute.manager [None req-92c34b61-f45d-4399-9e5e-4e6545f12f75 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Instance disappeared during snapshot {{(pid=61906) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 546.748375] env[61906]: INFO nova.compute.manager [-] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Took 1.05 seconds to deallocate network for instance. [ 546.752105] env[61906]: DEBUG nova.compute.claims [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 546.752307] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.804250] env[61906]: DEBUG nova.compute.manager [None req-92c34b61-f45d-4399-9e5e-4e6545f12f75 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Found 0 images (rotation: 2) {{(pid=61906) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 546.827560] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b3f4390-d58c-46ba-a943-14e838260ec5 req-3224b7ea-d9a7-41a4-abc5-cb58adee6174 service nova] Releasing lock "refresh_cache-f708ec4a-7f1e-4bfa-b561-4b9bc5556d27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.827836] env[61906]: DEBUG nova.compute.manager [req-5b3f4390-d58c-46ba-a943-14e838260ec5 req-3224b7ea-d9a7-41a4-abc5-cb58adee6174 service nova] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Received event network-vif-deleted-3cee9f61-8537-4ca1-bef4-683dfe02020c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 546.876666] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 05cefc82-3585-4d0b-8620-91412b3ef69e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 546.980878] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Releasing lock "refresh_cache-e23fd777-66f7-4ee1-bdcb-fe19c55f955c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.980878] env[61906]: DEBUG nova.compute.manager [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 546.980878] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 546.980878] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a492e1c5-7139-4b00-a999-17d078c96e99 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.992282] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 546.992282] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f5e3b7c-69e2-4db2-9368-457cb530e572 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.999251] env[61906]: DEBUG oslo_vmware.api [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Waiting for the task: (returnval){ [ 546.999251] env[61906]: value = "task-1356484" [ 546.999251] env[61906]: _type = "Task" [ 546.999251] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.009811] env[61906]: DEBUG oslo_vmware.api [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356484, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.382167] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 40f3efa6-699f-4779-b999-c86435e04c0a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 547.516175] env[61906]: DEBUG oslo_vmware.api [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356484, 'name': PowerOffVM_Task, 'duration_secs': 0.200285} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.516175] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 547.516175] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 547.516175] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fae2b600-ea58-4223-88f5-fe8a3a486f20 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.543635] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 547.543873] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 547.544070] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Deleting the datastore file [datastore2] e23fd777-66f7-4ee1-bdcb-fe19c55f955c {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 547.544332] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af4af179-e1f5-4a2c-9b08-dd988ac6b154 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.554409] env[61906]: DEBUG oslo_vmware.api [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Waiting for the task: (returnval){ [ 547.554409] env[61906]: value = "task-1356486" [ 547.554409] env[61906]: _type = "Task" [ 547.554409] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.565519] env[61906]: DEBUG oslo_vmware.api [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356486, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.645511] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7371668-f2ba-4576-8b5e-e2856fae326f tempest-ServersListShow296Test-1608988043 tempest-ServersListShow296Test-1608988043-project-member] Acquiring lock "972379ac-0384-4459-8da4-54d747d1f946" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.645777] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7371668-f2ba-4576-8b5e-e2856fae326f tempest-ServersListShow296Test-1608988043 tempest-ServersListShow296Test-1608988043-project-member] Lock "972379ac-0384-4459-8da4-54d747d1f946" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.886275] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 081d06b3-c2b6-42f2-94b9-925493590e55 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 548.067160] env[61906]: DEBUG oslo_vmware.api [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Task: {'id': task-1356486, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.111477} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.068015] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 548.068215] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 548.068397] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 548.068581] env[61906]: INFO nova.compute.manager [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Took 1.09 seconds to destroy the instance on the hypervisor. [ 548.068820] env[61906]: DEBUG oslo.service.loopingcall [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.069021] env[61906]: DEBUG nova.compute.manager [-] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.069112] env[61906]: DEBUG nova.network.neutron [-] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 548.089583] env[61906]: DEBUG nova.network.neutron [-] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.304127] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Acquiring lock "71fe23a8-1108-4781-9a35-0156cb9d584d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.304510] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Lock "71fe23a8-1108-4781-9a35-0156cb9d584d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.392504] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance ebea667f-6495-41c0-88b3-a9c61390558d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 548.592088] env[61906]: DEBUG nova.network.neutron [-] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.791506] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquiring lock "8d4c25cc-47d0-42e0-8fe9-49426175af68" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.791736] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "8d4c25cc-47d0-42e0-8fe9-49426175af68" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.899872] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance b56dd330-d31c-4e42-baf9-d9897c55e633 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 549.097849] env[61906]: INFO nova.compute.manager [-] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Took 1.03 seconds to deallocate network for instance. [ 549.403473] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance f9edb738-210e-4f85-8062-7759fa9033ea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 549.604804] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.907595] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 3e6732aa-a878-4bde-9299-abd779f2a109 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 550.410992] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 73a96c0e-a08b-46c9-b700-25544e17c73d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 550.918535] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 0c7d1af9-54e9-435d-a0cc-687f793d4f43 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 551.421267] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 7aee8304-8d07-4aa2-867a-2eeeeacbabc8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 551.924862] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance e8a2267e-ae4e-4f73-9691-3163ba3ddaa7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 552.427638] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 569d7ab1-22b8-4006-bf0e-ff35032cb9a6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 552.930793] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance e5c81f8b-aac3-4cf8-9abb-444731f57206 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 553.433504] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 95604a75-0ab3-45d9-a235-9067e615e72a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 553.433807] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 553.433903] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 553.777494] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2adfa5d7-466c-4205-a40a-afc506eb9a64 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.785809] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d41753-28b8-4091-8d4f-d3ad421d5066 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.814890] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9151620b-f1fe-426b-892e-0181c1a97215 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.822690] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2777f484-08e6-47a1-af72-d7aee4c7a20f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.835786] env[61906]: DEBUG nova.compute.provider_tree [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 554.339028] env[61906]: DEBUG nova.scheduler.client.report [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 554.844498] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 554.844776] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.080s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.845939] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 32.370s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.847895] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.848272] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Getting list of instances from cluster (obj){ [ 554.848272] env[61906]: value = "domain-c8" [ 554.848272] env[61906]: _type = "ClusterComputeResource" [ 554.848272] env[61906]: } {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 554.849463] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f7f9b9-112f-4b28-9f99-d7545e85bdcd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.859201] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Got total of 0 instances {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 555.691328] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4436d917-cc86-455d-82d2-5f85cdac7a6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.699844] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1732daf5-fd51-48c7-9680-c0d9bb9d2144 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.729301] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5468ed-6ad7-4959-85d1-773e3f5aaea4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.742903] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37073a0b-de62-47a5-87c2-2690a9a8aa2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.756760] env[61906]: DEBUG nova.compute.provider_tree [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.260375] env[61906]: DEBUG nova.scheduler.client.report [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 556.767138] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.922s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.767912] env[61906]: ERROR nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 74e04542-2f68-486a-8d7c-48ae3f35e469, please check neutron logs for more information. [ 556.767912] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Traceback (most recent call last): [ 556.767912] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.767912] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] self.driver.spawn(context, instance, image_meta, [ 556.767912] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 556.767912] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.767912] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.767912] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] vm_ref = self.build_virtual_machine(instance, [ 556.767912] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.767912] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.767912] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.768397] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] for vif in network_info: [ 556.768397] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.768397] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] return self._sync_wrapper(fn, *args, **kwargs) [ 556.768397] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.768397] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] self.wait() [ 556.768397] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.768397] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] self[:] = self._gt.wait() [ 556.768397] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.768397] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] return self._exit_event.wait() [ 556.768397] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.768397] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] result = hub.switch() [ 556.768397] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.768397] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] return self.greenlet.switch() [ 556.768752] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.768752] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] result = function(*args, **kwargs) [ 556.768752] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 556.768752] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] return func(*args, **kwargs) [ 556.768752] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.768752] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] raise e [ 556.768752] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.768752] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] nwinfo = self.network_api.allocate_for_instance( [ 556.768752] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.768752] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] created_port_ids = self._update_ports_for_instance( [ 556.768752] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.768752] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] with excutils.save_and_reraise_exception(): [ 556.768752] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.769113] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] self.force_reraise() [ 556.769113] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.769113] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] raise self.value [ 556.769113] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.769113] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] updated_port = self._update_port( [ 556.769113] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.769113] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] _ensure_no_port_binding_failure(port) [ 556.769113] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.769113] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] raise exception.PortBindingFailed(port_id=port['id']) [ 556.769113] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] nova.exception.PortBindingFailed: Binding failed for port 74e04542-2f68-486a-8d7c-48ae3f35e469, please check neutron logs for more information. [ 556.769113] env[61906]: ERROR nova.compute.manager [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] [ 556.769428] env[61906]: DEBUG nova.compute.utils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Binding failed for port 74e04542-2f68-486a-8d7c-48ae3f35e469, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 556.770175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.662s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.771795] env[61906]: INFO nova.compute.claims [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 556.775238] env[61906]: DEBUG nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Build of instance fa0cbe6c-937b-4276-a419-0ab6e92e795f was re-scheduled: Binding failed for port 74e04542-2f68-486a-8d7c-48ae3f35e469, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 556.775751] env[61906]: DEBUG nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 556.776032] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Acquiring lock "refresh_cache-fa0cbe6c-937b-4276-a419-0ab6e92e795f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.776258] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Acquired lock "refresh_cache-fa0cbe6c-937b-4276-a419-0ab6e92e795f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.776472] env[61906]: DEBUG nova.network.neutron [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 557.306112] env[61906]: DEBUG nova.network.neutron [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.371485] env[61906]: DEBUG nova.network.neutron [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.875279] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Releasing lock "refresh_cache-fa0cbe6c-937b-4276-a419-0ab6e92e795f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.875966] env[61906]: DEBUG nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 557.875966] env[61906]: DEBUG nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.875966] env[61906]: DEBUG nova.network.neutron [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 557.898465] env[61906]: DEBUG nova.network.neutron [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.155388] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceed382b-a306-49af-a612-0475a3bc9ef5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.163089] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca84181-22af-4fee-a6fd-79cf44cb1774 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.194461] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5693bb30-ac48-4d39-9ad7-b39d511b7776 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.205029] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b31d734-e69a-4e5c-8fa5-d9e69a668659 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.216975] env[61906]: DEBUG nova.compute.provider_tree [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.400803] env[61906]: DEBUG nova.network.neutron [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.721057] env[61906]: DEBUG nova.scheduler.client.report [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 558.903143] env[61906]: INFO nova.compute.manager [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] [instance: fa0cbe6c-937b-4276-a419-0ab6e92e795f] Took 1.03 seconds to deallocate network for instance. [ 559.226048] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.227910] env[61906]: DEBUG nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 559.233055] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 33.501s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.738858] env[61906]: DEBUG nova.compute.utils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 559.740079] env[61906]: DEBUG nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 559.740253] env[61906]: DEBUG nova.network.neutron [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 559.789292] env[61906]: DEBUG nova.policy [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7a99fe7339084934998328a2a626dd08', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0495675ff94f494aa8ab2cc86e8c74d3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 559.930830] env[61906]: INFO nova.scheduler.client.report [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Deleted allocations for instance fa0cbe6c-937b-4276-a419-0ab6e92e795f [ 560.137657] env[61906]: DEBUG nova.network.neutron [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Successfully created port: 009e2ac7-b58e-4c4b-ba55-9b8ccda347fe {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 560.175407] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5fada33-5d86-481b-ab53-57f175f5a174 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.184612] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6615f5aa-5773-466d-b5ac-acf45c8f3d53 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.213931] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986babb6-5e36-4402-bbb6-2d223879b00e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.221403] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4f8afed-c131-4517-a65f-3eb09a1ab233 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.234982] env[61906]: DEBUG nova.compute.provider_tree [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.243156] env[61906]: DEBUG nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 560.444548] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4edf465d-12f3-439f-8663-ba42489dc3fa tempest-ServersAdminNegativeTestJSON-1886528465 tempest-ServersAdminNegativeTestJSON-1886528465-project-member] Lock "fa0cbe6c-937b-4276-a419-0ab6e92e795f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.893s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.742308] env[61906]: DEBUG nova.scheduler.client.report [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 560.949301] env[61906]: DEBUG nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.950031] env[61906]: DEBUG nova.compute.manager [req-6d7125f0-6625-4146-b501-a0cd4b57c0a5 req-eb8a5e34-ece8-4aa6-9509-33f866c47a3c service nova] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Received event network-changed-009e2ac7-b58e-4c4b-ba55-9b8ccda347fe {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.950313] env[61906]: DEBUG nova.compute.manager [req-6d7125f0-6625-4146-b501-a0cd4b57c0a5 req-eb8a5e34-ece8-4aa6-9509-33f866c47a3c service nova] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Refreshing instance network info cache due to event network-changed-009e2ac7-b58e-4c4b-ba55-9b8ccda347fe. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 560.950546] env[61906]: DEBUG oslo_concurrency.lockutils [req-6d7125f0-6625-4146-b501-a0cd4b57c0a5 req-eb8a5e34-ece8-4aa6-9509-33f866c47a3c service nova] Acquiring lock "refresh_cache-ec81c4ec-07c6-4b1f-8555-69902cff4765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.951043] env[61906]: DEBUG oslo_concurrency.lockutils [req-6d7125f0-6625-4146-b501-a0cd4b57c0a5 req-eb8a5e34-ece8-4aa6-9509-33f866c47a3c service nova] Acquired lock "refresh_cache-ec81c4ec-07c6-4b1f-8555-69902cff4765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.951249] env[61906]: DEBUG nova.network.neutron [req-6d7125f0-6625-4146-b501-a0cd4b57c0a5 req-eb8a5e34-ece8-4aa6-9509-33f866c47a3c service nova] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Refreshing network info cache for port 009e2ac7-b58e-4c4b-ba55-9b8ccda347fe {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 561.247499] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.017s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.248149] env[61906]: ERROR nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 65707ace-cf18-4ce6-8f31-6802abce3c29, please check neutron logs for more information. [ 561.248149] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Traceback (most recent call last): [ 561.248149] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 561.248149] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] self.driver.spawn(context, instance, image_meta, [ 561.248149] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 561.248149] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.248149] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.248149] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] vm_ref = self.build_virtual_machine(instance, [ 561.248149] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.248149] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.248149] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.248529] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] for vif in network_info: [ 561.248529] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 561.248529] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] return self._sync_wrapper(fn, *args, **kwargs) [ 561.248529] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 561.248529] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] self.wait() [ 561.248529] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 561.248529] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] self[:] = self._gt.wait() [ 561.248529] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.248529] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] return self._exit_event.wait() [ 561.248529] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.248529] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] result = hub.switch() [ 561.248529] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.248529] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] return self.greenlet.switch() [ 561.248921] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.248921] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] result = function(*args, **kwargs) [ 561.248921] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 561.248921] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] return func(*args, **kwargs) [ 561.248921] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.248921] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] raise e [ 561.248921] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.248921] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] nwinfo = self.network_api.allocate_for_instance( [ 561.248921] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.248921] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] created_port_ids = self._update_ports_for_instance( [ 561.248921] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.248921] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] with excutils.save_and_reraise_exception(): [ 561.248921] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.249315] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] self.force_reraise() [ 561.249315] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.249315] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] raise self.value [ 561.249315] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.249315] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] updated_port = self._update_port( [ 561.249315] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.249315] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] _ensure_no_port_binding_failure(port) [ 561.249315] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.249315] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] raise exception.PortBindingFailed(port_id=port['id']) [ 561.249315] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] nova.exception.PortBindingFailed: Binding failed for port 65707ace-cf18-4ce6-8f31-6802abce3c29, please check neutron logs for more information. [ 561.249315] env[61906]: ERROR nova.compute.manager [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] [ 561.249649] env[61906]: DEBUG nova.compute.utils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Binding failed for port 65707ace-cf18-4ce6-8f31-6802abce3c29, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 561.250375] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 32.389s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.262024] env[61906]: DEBUG nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Build of instance 9f6864a2-3e54-41cb-b858-16e6e350a2af was re-scheduled: Binding failed for port 65707ace-cf18-4ce6-8f31-6802abce3c29, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 561.262024] env[61906]: DEBUG nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 561.262024] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Acquiring lock "refresh_cache-9f6864a2-3e54-41cb-b858-16e6e350a2af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.262024] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Acquired lock "refresh_cache-9f6864a2-3e54-41cb-b858-16e6e350a2af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.262368] env[61906]: DEBUG nova.network.neutron [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 561.262368] env[61906]: DEBUG nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 561.306915] env[61906]: DEBUG nova.virt.hardware [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 561.306915] env[61906]: DEBUG nova.virt.hardware [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 561.306915] env[61906]: DEBUG nova.virt.hardware [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 561.306915] env[61906]: DEBUG nova.virt.hardware [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 561.307392] env[61906]: DEBUG nova.virt.hardware [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 561.307392] env[61906]: DEBUG nova.virt.hardware [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 561.307752] env[61906]: DEBUG nova.virt.hardware [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 561.307752] env[61906]: DEBUG nova.virt.hardware [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 561.308397] env[61906]: DEBUG nova.virt.hardware [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 561.308397] env[61906]: DEBUG nova.virt.hardware [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 561.308570] env[61906]: DEBUG nova.virt.hardware [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 561.311575] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c60dc7a-5c81-4147-9b84-14bd7495ab4b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.324679] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064e1a9a-05f9-4ae3-82ee-6eade3e6f59d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.475686] env[61906]: ERROR nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 009e2ac7-b58e-4c4b-ba55-9b8ccda347fe, please check neutron logs for more information. [ 561.475686] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 561.475686] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.475686] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 561.475686] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.475686] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 561.475686] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.475686] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 561.475686] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.475686] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 561.475686] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.475686] env[61906]: ERROR nova.compute.manager raise self.value [ 561.475686] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.475686] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 561.475686] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.475686] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 561.476223] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.476223] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 561.476223] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 009e2ac7-b58e-4c4b-ba55-9b8ccda347fe, please check neutron logs for more information. [ 561.476223] env[61906]: ERROR nova.compute.manager [ 561.476223] env[61906]: Traceback (most recent call last): [ 561.476223] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 561.476223] env[61906]: listener.cb(fileno) [ 561.476223] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.476223] env[61906]: result = function(*args, **kwargs) [ 561.476223] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 561.476223] env[61906]: return func(*args, **kwargs) [ 561.476223] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.476223] env[61906]: raise e [ 561.476223] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.476223] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 561.476223] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.476223] env[61906]: created_port_ids = self._update_ports_for_instance( [ 561.476223] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.476223] env[61906]: with excutils.save_and_reraise_exception(): [ 561.476223] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.476223] env[61906]: self.force_reraise() [ 561.476223] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.476223] env[61906]: raise self.value [ 561.476223] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.476223] env[61906]: updated_port = self._update_port( [ 561.476223] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.476223] env[61906]: _ensure_no_port_binding_failure(port) [ 561.476223] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.476223] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 561.477229] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 009e2ac7-b58e-4c4b-ba55-9b8ccda347fe, please check neutron logs for more information. [ 561.477229] env[61906]: Removing descriptor: 17 [ 561.477229] env[61906]: ERROR nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 009e2ac7-b58e-4c4b-ba55-9b8ccda347fe, please check neutron logs for more information. [ 561.477229] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Traceback (most recent call last): [ 561.477229] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 561.477229] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] yield resources [ 561.477229] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 561.477229] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] self.driver.spawn(context, instance, image_meta, [ 561.477229] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 561.477229] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.477229] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.477229] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] vm_ref = self.build_virtual_machine(instance, [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] for vif in network_info: [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] return self._sync_wrapper(fn, *args, **kwargs) [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] self.wait() [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] self[:] = self._gt.wait() [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] return self._exit_event.wait() [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.477579] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] result = hub.switch() [ 561.477971] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.477971] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] return self.greenlet.switch() [ 561.477971] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.477971] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] result = function(*args, **kwargs) [ 561.477971] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 561.477971] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] return func(*args, **kwargs) [ 561.477971] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.477971] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] raise e [ 561.477971] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.477971] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] nwinfo = self.network_api.allocate_for_instance( [ 561.477971] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.477971] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] created_port_ids = self._update_ports_for_instance( [ 561.477971] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.478368] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] with excutils.save_and_reraise_exception(): [ 561.478368] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.478368] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] self.force_reraise() [ 561.478368] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.478368] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] raise self.value [ 561.478368] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.478368] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] updated_port = self._update_port( [ 561.478368] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.478368] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] _ensure_no_port_binding_failure(port) [ 561.478368] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.478368] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] raise exception.PortBindingFailed(port_id=port['id']) [ 561.478368] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] nova.exception.PortBindingFailed: Binding failed for port 009e2ac7-b58e-4c4b-ba55-9b8ccda347fe, please check neutron logs for more information. [ 561.478368] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] [ 561.478756] env[61906]: INFO nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Terminating instance [ 561.480732] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.481017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquiring lock "refresh_cache-ec81c4ec-07c6-4b1f-8555-69902cff4765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.482600] env[61906]: DEBUG nova.network.neutron [req-6d7125f0-6625-4146-b501-a0cd4b57c0a5 req-eb8a5e34-ece8-4aa6-9509-33f866c47a3c service nova] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.583031] env[61906]: DEBUG nova.network.neutron [req-6d7125f0-6625-4146-b501-a0cd4b57c0a5 req-eb8a5e34-ece8-4aa6-9509-33f866c47a3c service nova] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.781229] env[61906]: DEBUG nova.network.neutron [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.867318] env[61906]: DEBUG nova.network.neutron [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.085071] env[61906]: DEBUG oslo_concurrency.lockutils [req-6d7125f0-6625-4146-b501-a0cd4b57c0a5 req-eb8a5e34-ece8-4aa6-9509-33f866c47a3c service nova] Releasing lock "refresh_cache-ec81c4ec-07c6-4b1f-8555-69902cff4765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.085442] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquired lock "refresh_cache-ec81c4ec-07c6-4b1f-8555-69902cff4765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.085628] env[61906]: DEBUG nova.network.neutron [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 562.177742] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0ba961-7627-4321-9d5f-30ae354954d6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.185803] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0093c765-d0c1-4b4b-8356-f92b1ec9ebe3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.215375] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c2b53e8-69c7-4266-a6bb-88e5d16ca6d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.223134] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb478d50-a77f-4e6a-9d74-2aac77910f69 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.236627] env[61906]: DEBUG nova.compute.provider_tree [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.372508] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Releasing lock "refresh_cache-9f6864a2-3e54-41cb-b858-16e6e350a2af" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.372766] env[61906]: DEBUG nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 562.372951] env[61906]: DEBUG nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.373142] env[61906]: DEBUG nova.network.neutron [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 562.387484] env[61906]: DEBUG nova.network.neutron [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.603064] env[61906]: DEBUG nova.network.neutron [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.698411] env[61906]: DEBUG nova.network.neutron [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.739931] env[61906]: DEBUG nova.scheduler.client.report [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 562.891571] env[61906]: DEBUG nova.network.neutron [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.142532] env[61906]: DEBUG nova.compute.manager [req-ad2a59c5-f9eb-47a3-852c-6ffbb0e030b9 req-972dc9a8-0d0c-4359-a2a5-992d552cc453 service nova] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Received event network-vif-deleted-009e2ac7-b58e-4c4b-ba55-9b8ccda347fe {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 563.201319] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Releasing lock "refresh_cache-ec81c4ec-07c6-4b1f-8555-69902cff4765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.201754] env[61906]: DEBUG nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 563.201942] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 563.202625] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bfd5a0ff-44fc-41f1-9875-98f37bafeafa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.212763] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fda90f1-f583-4501-9c30-7fe1a80240af {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.236610] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ec81c4ec-07c6-4b1f-8555-69902cff4765 could not be found. [ 563.236860] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 563.237087] env[61906]: INFO nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Took 0.04 seconds to destroy the instance on the hypervisor. [ 563.237364] env[61906]: DEBUG oslo.service.loopingcall [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 563.237614] env[61906]: DEBUG nova.compute.manager [-] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.237707] env[61906]: DEBUG nova.network.neutron [-] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 563.245476] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.246035] env[61906]: ERROR nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7caa7da5-777c-43bf-8628-f8741f621e94, please check neutron logs for more information. [ 563.246035] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Traceback (most recent call last): [ 563.246035] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 563.246035] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] self.driver.spawn(context, instance, image_meta, [ 563.246035] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 563.246035] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.246035] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.246035] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] vm_ref = self.build_virtual_machine(instance, [ 563.246035] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.246035] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.246035] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.246548] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] for vif in network_info: [ 563.246548] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 563.246548] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] return self._sync_wrapper(fn, *args, **kwargs) [ 563.246548] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 563.246548] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] self.wait() [ 563.246548] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 563.246548] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] self[:] = self._gt.wait() [ 563.246548] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.246548] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] return self._exit_event.wait() [ 563.246548] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.246548] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] result = hub.switch() [ 563.246548] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.246548] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] return self.greenlet.switch() [ 563.246921] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.246921] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] result = function(*args, **kwargs) [ 563.246921] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 563.246921] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] return func(*args, **kwargs) [ 563.246921] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.246921] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] raise e [ 563.246921] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.246921] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] nwinfo = self.network_api.allocate_for_instance( [ 563.246921] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.246921] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] created_port_ids = self._update_ports_for_instance( [ 563.246921] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.246921] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] with excutils.save_and_reraise_exception(): [ 563.246921] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.247296] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] self.force_reraise() [ 563.247296] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.247296] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] raise self.value [ 563.247296] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.247296] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] updated_port = self._update_port( [ 563.247296] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.247296] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] _ensure_no_port_binding_failure(port) [ 563.247296] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.247296] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] raise exception.PortBindingFailed(port_id=port['id']) [ 563.247296] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] nova.exception.PortBindingFailed: Binding failed for port 7caa7da5-777c-43bf-8628-f8741f621e94, please check neutron logs for more information. [ 563.247296] env[61906]: ERROR nova.compute.manager [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] [ 563.247597] env[61906]: DEBUG nova.compute.utils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Binding failed for port 7caa7da5-777c-43bf-8628-f8741f621e94, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 563.247804] env[61906]: DEBUG oslo_concurrency.lockutils [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 34.287s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.247976] env[61906]: DEBUG nova.objects.instance [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61906) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 563.250688] env[61906]: DEBUG nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Build of instance 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db was re-scheduled: Binding failed for port 7caa7da5-777c-43bf-8628-f8741f621e94, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 563.251836] env[61906]: DEBUG nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 563.251836] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Acquiring lock "refresh_cache-4dc3af05-b670-4f7c-bb9f-d1fa2e4180db" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.251836] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Acquired lock "refresh_cache-4dc3af05-b670-4f7c-bb9f-d1fa2e4180db" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.251836] env[61906]: DEBUG nova.network.neutron [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 563.262760] env[61906]: DEBUG nova.network.neutron [-] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.393439] env[61906]: INFO nova.compute.manager [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] [instance: 9f6864a2-3e54-41cb-b858-16e6e350a2af] Took 1.02 seconds to deallocate network for instance. [ 563.763125] env[61906]: DEBUG nova.network.neutron [-] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.772350] env[61906]: DEBUG nova.network.neutron [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.858490] env[61906]: DEBUG nova.network.neutron [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.264078] env[61906]: DEBUG oslo_concurrency.lockutils [None req-36fd4f58-513e-488f-b972-1ccca40c271a tempest-ServersAdmin275Test-17191007 tempest-ServersAdmin275Test-17191007-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.265244] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.262s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.267997] env[61906]: INFO nova.compute.claims [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 564.277201] env[61906]: INFO nova.compute.manager [-] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Took 1.04 seconds to deallocate network for instance. [ 564.282383] env[61906]: DEBUG nova.compute.claims [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 564.282571] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.362038] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Releasing lock "refresh_cache-4dc3af05-b670-4f7c-bb9f-d1fa2e4180db" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.362038] env[61906]: DEBUG nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 564.362038] env[61906]: DEBUG nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 564.362038] env[61906]: DEBUG nova.network.neutron [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.377924] env[61906]: DEBUG nova.network.neutron [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.434210] env[61906]: INFO nova.scheduler.client.report [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Deleted allocations for instance 9f6864a2-3e54-41cb-b858-16e6e350a2af [ 564.881052] env[61906]: DEBUG nova.network.neutron [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.942927] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cb11b1b0-a1fc-4933-a2a7-f60ea8584e28 tempest-ServersWithSpecificFlavorTestJSON-1138552807 tempest-ServersWithSpecificFlavorTestJSON-1138552807-project-member] Lock "9f6864a2-3e54-41cb-b858-16e6e350a2af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.924s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.386433] env[61906]: INFO nova.compute.manager [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] [instance: 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db] Took 1.02 seconds to deallocate network for instance. [ 565.447745] env[61906]: DEBUG nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 565.738510] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aac22fb6-fb08-4ba3-bac7-ce7d77cba512 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.751022] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0032b3a-e2dd-48cf-bf3f-a5803e8556fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.784202] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3676a6-dae0-4fd8-8923-0055efde32f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.791422] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26a163c-6557-4616-8098-648ec2ffe09c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.807205] env[61906]: DEBUG nova.compute.provider_tree [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.977255] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.311641] env[61906]: DEBUG nova.scheduler.client.report [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.336930] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Acquiring lock "2f5bd366-317e-4c18-956a-35d9cdbae9e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.337149] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Lock "2f5bd366-317e-4c18-956a-35d9cdbae9e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.420367] env[61906]: INFO nova.scheduler.client.report [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Deleted allocations for instance 4dc3af05-b670-4f7c-bb9f-d1fa2e4180db [ 566.817683] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.552s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.818196] env[61906]: DEBUG nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 566.825315] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.459s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.825513] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.827528] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.092s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.829032] env[61906]: INFO nova.compute.claims [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 566.859967] env[61906]: INFO nova.scheduler.client.report [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Deleted allocations for instance 85effb81-494c-4bda-a093-6dde523d5f45 [ 566.928369] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3ea528d7-a4d7-4842-8304-7e4e53c40ef9 tempest-VolumesAssistedSnapshotsTest-804680528 tempest-VolumesAssistedSnapshotsTest-804680528-project-member] Lock "4dc3af05-b670-4f7c-bb9f-d1fa2e4180db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.932s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.340019] env[61906]: DEBUG nova.compute.utils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 567.342356] env[61906]: DEBUG nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 567.342665] env[61906]: DEBUG nova.network.neutron [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 567.373384] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bafd0c31-d219-472d-b2dd-7a2df81f6496 tempest-ServersAdmin275Test-2024162162 tempest-ServersAdmin275Test-2024162162-project-member] Lock "85effb81-494c-4bda-a093-6dde523d5f45" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.791s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.432897] env[61906]: DEBUG nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 567.438487] env[61906]: DEBUG nova.policy [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7a99fe7339084934998328a2a626dd08', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0495675ff94f494aa8ab2cc86e8c74d3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 567.843464] env[61906]: DEBUG nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 567.954706] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.158456] env[61906]: DEBUG nova.network.neutron [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Successfully created port: faa99eaa-4906-4750-b710-7771eb7cf572 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 568.291043] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0433f9c9-9c40-49fe-aa2f-c1c230929e8d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.300454] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ec274a-9e83-44a4-bfc3-bcc979700712 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.337465] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa96128-3ac6-4063-b2c8-07fd6f14479e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.346274] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-204d6eff-c827-47a6-b9c6-66cb66b4e7bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.364651] env[61906]: DEBUG nova.compute.provider_tree [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 568.868508] env[61906]: DEBUG nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 568.875799] env[61906]: DEBUG nova.scheduler.client.report [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 568.906062] env[61906]: DEBUG nova.virt.hardware [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 568.906062] env[61906]: DEBUG nova.virt.hardware [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 568.906062] env[61906]: DEBUG nova.virt.hardware [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 568.906062] env[61906]: DEBUG nova.virt.hardware [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 568.906340] env[61906]: DEBUG nova.virt.hardware [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 568.906340] env[61906]: DEBUG nova.virt.hardware [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 568.906340] env[61906]: DEBUG nova.virt.hardware [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 568.906340] env[61906]: DEBUG nova.virt.hardware [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 568.906340] env[61906]: DEBUG nova.virt.hardware [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 568.906529] env[61906]: DEBUG nova.virt.hardware [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 568.906681] env[61906]: DEBUG nova.virt.hardware [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 568.908452] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a715c725-cac9-43f1-b227-64dde0352b13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.916857] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4898f07f-ec1d-4c76-901b-2266ccb282d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.379883] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.552s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.380396] env[61906]: DEBUG nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 569.387022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.917s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.387022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.387022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.095s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.389151] env[61906]: INFO nova.compute.claims [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 569.442996] env[61906]: INFO nova.scheduler.client.report [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Deleted allocations for instance 7f4bb9fb-fa64-451d-9876-7e4314c13a64 [ 569.889336] env[61906]: DEBUG nova.compute.manager [req-f341f435-3f9c-4501-baa0-474b2d928544 req-6718edbc-60af-4626-a781-873fe08064da service nova] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Received event network-changed-faa99eaa-4906-4750-b710-7771eb7cf572 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 569.893379] env[61906]: DEBUG nova.compute.manager [req-f341f435-3f9c-4501-baa0-474b2d928544 req-6718edbc-60af-4626-a781-873fe08064da service nova] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Refreshing instance network info cache due to event network-changed-faa99eaa-4906-4750-b710-7771eb7cf572. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 569.893643] env[61906]: DEBUG oslo_concurrency.lockutils [req-f341f435-3f9c-4501-baa0-474b2d928544 req-6718edbc-60af-4626-a781-873fe08064da service nova] Acquiring lock "refresh_cache-f72e6538-7130-45a9-a722-ba60de3bda49" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.893803] env[61906]: DEBUG oslo_concurrency.lockutils [req-f341f435-3f9c-4501-baa0-474b2d928544 req-6718edbc-60af-4626-a781-873fe08064da service nova] Acquired lock "refresh_cache-f72e6538-7130-45a9-a722-ba60de3bda49" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.894035] env[61906]: DEBUG nova.network.neutron [req-f341f435-3f9c-4501-baa0-474b2d928544 req-6718edbc-60af-4626-a781-873fe08064da service nova] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Refreshing network info cache for port faa99eaa-4906-4750-b710-7771eb7cf572 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 569.896483] env[61906]: DEBUG nova.compute.utils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 569.903116] env[61906]: DEBUG nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 569.903257] env[61906]: DEBUG nova.network.neutron [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 569.956761] env[61906]: DEBUG oslo_concurrency.lockutils [None req-638f5759-e359-4dc6-9439-4e0a48983007 tempest-ServerDiagnosticsV248Test-1613281579 tempest-ServerDiagnosticsV248Test-1613281579-project-member] Lock "7f4bb9fb-fa64-451d-9876-7e4314c13a64" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.449s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.000107] env[61906]: DEBUG nova.policy [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a4f6272a837477aa04bc88b55ec857d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9db10096c0254e258125ba7c333c8161', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 570.405021] env[61906]: DEBUG nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 570.500092] env[61906]: DEBUG nova.network.neutron [req-f341f435-3f9c-4501-baa0-474b2d928544 req-6718edbc-60af-4626-a781-873fe08064da service nova] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.658124] env[61906]: ERROR nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port faa99eaa-4906-4750-b710-7771eb7cf572, please check neutron logs for more information. [ 570.658124] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 570.658124] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 570.658124] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 570.658124] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 570.658124] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 570.658124] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 570.658124] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 570.658124] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 570.658124] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 570.658124] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 570.658124] env[61906]: ERROR nova.compute.manager raise self.value [ 570.658124] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 570.658124] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 570.658124] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 570.658124] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 570.658744] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 570.658744] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 570.658744] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port faa99eaa-4906-4750-b710-7771eb7cf572, please check neutron logs for more information. [ 570.658744] env[61906]: ERROR nova.compute.manager [ 570.658744] env[61906]: Traceback (most recent call last): [ 570.658744] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 570.658744] env[61906]: listener.cb(fileno) [ 570.658744] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 570.658744] env[61906]: result = function(*args, **kwargs) [ 570.658744] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 570.658744] env[61906]: return func(*args, **kwargs) [ 570.658744] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 570.658744] env[61906]: raise e [ 570.658744] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 570.658744] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 570.658744] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 570.658744] env[61906]: created_port_ids = self._update_ports_for_instance( [ 570.658744] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 570.658744] env[61906]: with excutils.save_and_reraise_exception(): [ 570.658744] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 570.658744] env[61906]: self.force_reraise() [ 570.658744] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 570.658744] env[61906]: raise self.value [ 570.658744] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 570.658744] env[61906]: updated_port = self._update_port( [ 570.658744] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 570.658744] env[61906]: _ensure_no_port_binding_failure(port) [ 570.658744] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 570.658744] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 570.659675] env[61906]: nova.exception.PortBindingFailed: Binding failed for port faa99eaa-4906-4750-b710-7771eb7cf572, please check neutron logs for more information. [ 570.659675] env[61906]: Removing descriptor: 16 [ 570.659675] env[61906]: ERROR nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port faa99eaa-4906-4750-b710-7771eb7cf572, please check neutron logs for more information. [ 570.659675] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Traceback (most recent call last): [ 570.659675] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 570.659675] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] yield resources [ 570.659675] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 570.659675] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] self.driver.spawn(context, instance, image_meta, [ 570.659675] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 570.659675] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] self._vmops.spawn(context, instance, image_meta, injected_files, [ 570.659675] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 570.659675] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] vm_ref = self.build_virtual_machine(instance, [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] vif_infos = vmwarevif.get_vif_info(self._session, [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] for vif in network_info: [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] return self._sync_wrapper(fn, *args, **kwargs) [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] self.wait() [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] self[:] = self._gt.wait() [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] return self._exit_event.wait() [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 570.660127] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] result = hub.switch() [ 570.660585] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 570.660585] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] return self.greenlet.switch() [ 570.660585] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 570.660585] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] result = function(*args, **kwargs) [ 570.660585] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 570.660585] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] return func(*args, **kwargs) [ 570.660585] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 570.660585] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] raise e [ 570.660585] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 570.660585] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] nwinfo = self.network_api.allocate_for_instance( [ 570.660585] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 570.660585] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] created_port_ids = self._update_ports_for_instance( [ 570.660585] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 570.660952] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] with excutils.save_and_reraise_exception(): [ 570.660952] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 570.660952] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] self.force_reraise() [ 570.660952] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 570.660952] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] raise self.value [ 570.660952] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 570.660952] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] updated_port = self._update_port( [ 570.660952] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 570.660952] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] _ensure_no_port_binding_failure(port) [ 570.660952] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 570.660952] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] raise exception.PortBindingFailed(port_id=port['id']) [ 570.660952] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] nova.exception.PortBindingFailed: Binding failed for port faa99eaa-4906-4750-b710-7771eb7cf572, please check neutron logs for more information. [ 570.660952] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] [ 570.661342] env[61906]: INFO nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Terminating instance [ 570.665177] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquiring lock "refresh_cache-f72e6538-7130-45a9-a722-ba60de3bda49" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.732019] env[61906]: DEBUG nova.network.neutron [req-f341f435-3f9c-4501-baa0-474b2d928544 req-6718edbc-60af-4626-a781-873fe08064da service nova] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.866024] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3346662b-ab9c-4966-be80-c1f527056698 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.871229] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7307d683-037c-439b-849a-bd97f3f4c8e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.905979] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bacc3fd9-d07d-4517-a0bf-b0c581c015d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.917556] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5af7ed9-4f14-4305-aea4-180308856c56 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.931381] env[61906]: DEBUG nova.compute.provider_tree [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.975959] env[61906]: DEBUG nova.network.neutron [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Successfully created port: c3930713-8fe6-470e-a2e4-6fe6c9377791 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 571.235260] env[61906]: DEBUG oslo_concurrency.lockutils [req-f341f435-3f9c-4501-baa0-474b2d928544 req-6718edbc-60af-4626-a781-873fe08064da service nova] Releasing lock "refresh_cache-f72e6538-7130-45a9-a722-ba60de3bda49" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.235746] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquired lock "refresh_cache-f72e6538-7130-45a9-a722-ba60de3bda49" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.236069] env[61906]: DEBUG nova.network.neutron [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 571.423427] env[61906]: DEBUG nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 571.435024] env[61906]: DEBUG nova.scheduler.client.report [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 571.463729] env[61906]: DEBUG nova.virt.hardware [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 571.464010] env[61906]: DEBUG nova.virt.hardware [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 571.464182] env[61906]: DEBUG nova.virt.hardware [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 571.464559] env[61906]: DEBUG nova.virt.hardware [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 571.464559] env[61906]: DEBUG nova.virt.hardware [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 571.464679] env[61906]: DEBUG nova.virt.hardware [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 571.464841] env[61906]: DEBUG nova.virt.hardware [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 571.465036] env[61906]: DEBUG nova.virt.hardware [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 571.465219] env[61906]: DEBUG nova.virt.hardware [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 571.465383] env[61906]: DEBUG nova.virt.hardware [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 571.465553] env[61906]: DEBUG nova.virt.hardware [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 571.466910] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-094235f3-3b64-4d2a-9418-b401ec60368e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.475585] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4176a9c2-0f32-4acd-852e-ea83d0e205ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.774168] env[61906]: DEBUG nova.network.neutron [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.941640] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.555s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.944138] env[61906]: DEBUG nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 571.946136] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.825s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.978282] env[61906]: DEBUG nova.network.neutron [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.990721] env[61906]: DEBUG nova.compute.manager [req-753a1320-346f-4234-a631-2b9082267fad req-34f3694d-e339-4871-913e-d70055b14461 service nova] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Received event network-vif-deleted-faa99eaa-4906-4750-b710-7771eb7cf572 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 572.452304] env[61906]: DEBUG nova.compute.utils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 572.460283] env[61906]: DEBUG nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 572.460467] env[61906]: DEBUG nova.network.neutron [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 572.481236] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Releasing lock "refresh_cache-f72e6538-7130-45a9-a722-ba60de3bda49" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.481650] env[61906]: DEBUG nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 572.481832] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 572.482159] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-738b33a5-12fe-4fb1-8327-3a4fd9162b9c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.496260] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76bf8fff-8124-470c-9a68-84ed2bb67a16 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.530271] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f72e6538-7130-45a9-a722-ba60de3bda49 could not be found. [ 572.531732] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 572.531732] env[61906]: INFO nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Took 0.05 seconds to destroy the instance on the hypervisor. [ 572.532038] env[61906]: DEBUG oslo.service.loopingcall [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 572.532424] env[61906]: DEBUG nova.compute.manager [-] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 572.532424] env[61906]: DEBUG nova.network.neutron [-] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 572.571641] env[61906]: DEBUG nova.network.neutron [-] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.649113] env[61906]: DEBUG nova.policy [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7175f91bc3b047a0b9db288897f38f63', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76581eb7c05947bab9ad415463cba670', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 572.752780] env[61906]: ERROR nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c3930713-8fe6-470e-a2e4-6fe6c9377791, please check neutron logs for more information. [ 572.752780] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 572.752780] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.752780] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 572.752780] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.752780] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 572.752780] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.752780] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 572.752780] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.752780] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 572.752780] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.752780] env[61906]: ERROR nova.compute.manager raise self.value [ 572.752780] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.752780] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 572.752780] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.752780] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 572.753316] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.753316] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 572.753316] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c3930713-8fe6-470e-a2e4-6fe6c9377791, please check neutron logs for more information. [ 572.753316] env[61906]: ERROR nova.compute.manager [ 572.753316] env[61906]: Traceback (most recent call last): [ 572.753316] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 572.753316] env[61906]: listener.cb(fileno) [ 572.753316] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.753316] env[61906]: result = function(*args, **kwargs) [ 572.753316] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 572.753316] env[61906]: return func(*args, **kwargs) [ 572.753316] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.753316] env[61906]: raise e [ 572.753316] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.753316] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 572.753316] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.753316] env[61906]: created_port_ids = self._update_ports_for_instance( [ 572.753316] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.753316] env[61906]: with excutils.save_and_reraise_exception(): [ 572.753316] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.753316] env[61906]: self.force_reraise() [ 572.753316] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.753316] env[61906]: raise self.value [ 572.753316] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.753316] env[61906]: updated_port = self._update_port( [ 572.753316] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.753316] env[61906]: _ensure_no_port_binding_failure(port) [ 572.753316] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.753316] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 572.754222] env[61906]: nova.exception.PortBindingFailed: Binding failed for port c3930713-8fe6-470e-a2e4-6fe6c9377791, please check neutron logs for more information. [ 572.754222] env[61906]: Removing descriptor: 17 [ 572.754222] env[61906]: ERROR nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c3930713-8fe6-470e-a2e4-6fe6c9377791, please check neutron logs for more information. [ 572.754222] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Traceback (most recent call last): [ 572.754222] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 572.754222] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] yield resources [ 572.754222] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.754222] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] self.driver.spawn(context, instance, image_meta, [ 572.754222] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 572.754222] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.754222] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.754222] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] vm_ref = self.build_virtual_machine(instance, [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] for vif in network_info: [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] return self._sync_wrapper(fn, *args, **kwargs) [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] self.wait() [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] self[:] = self._gt.wait() [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] return self._exit_event.wait() [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.754587] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] result = hub.switch() [ 572.755040] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.755040] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] return self.greenlet.switch() [ 572.755040] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.755040] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] result = function(*args, **kwargs) [ 572.755040] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 572.755040] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] return func(*args, **kwargs) [ 572.755040] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.755040] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] raise e [ 572.755040] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.755040] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] nwinfo = self.network_api.allocate_for_instance( [ 572.755040] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.755040] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] created_port_ids = self._update_ports_for_instance( [ 572.755040] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.755428] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] with excutils.save_and_reraise_exception(): [ 572.755428] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.755428] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] self.force_reraise() [ 572.755428] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.755428] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] raise self.value [ 572.755428] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.755428] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] updated_port = self._update_port( [ 572.755428] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.755428] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] _ensure_no_port_binding_failure(port) [ 572.755428] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.755428] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] raise exception.PortBindingFailed(port_id=port['id']) [ 572.755428] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] nova.exception.PortBindingFailed: Binding failed for port c3930713-8fe6-470e-a2e4-6fe6c9377791, please check neutron logs for more information. [ 572.755428] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] [ 572.755817] env[61906]: INFO nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Terminating instance [ 572.756482] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Acquiring lock "refresh_cache-6b3ecea4-fe20-4b6c-a987-59c20731b9ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.756482] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Acquired lock "refresh_cache-6b3ecea4-fe20-4b6c-a987-59c20731b9ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.756482] env[61906]: DEBUG nova.network.neutron [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.907242] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413476af-95c9-4f74-ad28-6aed76161f84 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.917065] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f37e0b-d26d-45cc-b301-fce45b4cccc6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.948723] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee61dea0-fcc2-4fa4-a067-a96c09df42de {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.956421] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10970b51-6099-4277-82c6-9c788371358e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.962056] env[61906]: DEBUG nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 572.974815] env[61906]: DEBUG nova.compute.provider_tree [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.075084] env[61906]: DEBUG nova.network.neutron [-] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.283043] env[61906]: DEBUG nova.network.neutron [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.428091] env[61906]: DEBUG nova.network.neutron [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.481719] env[61906]: DEBUG nova.scheduler.client.report [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 573.581543] env[61906]: INFO nova.compute.manager [-] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Took 1.05 seconds to deallocate network for instance. [ 573.583965] env[61906]: DEBUG nova.compute.claims [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 573.584182] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.757225] env[61906]: DEBUG nova.network.neutron [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Successfully created port: 208515e6-c9dd-4465-b0b9-0c97b49dab36 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 573.931357] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Releasing lock "refresh_cache-6b3ecea4-fe20-4b6c-a987-59c20731b9ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.932178] env[61906]: DEBUG nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 573.932364] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 573.932758] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6fa6076f-d6f4-4d70-b676-1e6e4b23443a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.946966] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fffbccc2-1b45-4946-bf37-e6aad477bc9a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.977090] env[61906]: DEBUG nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 573.981658] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6b3ecea4-fe20-4b6c-a987-59c20731b9ac could not be found. [ 573.981807] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 573.981993] env[61906]: INFO nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Took 0.05 seconds to destroy the instance on the hypervisor. [ 573.982263] env[61906]: DEBUG oslo.service.loopingcall [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 573.983726] env[61906]: DEBUG nova.compute.manager [-] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 573.983726] env[61906]: DEBUG nova.network.neutron [-] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 573.989698] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.044s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.990582] env[61906]: ERROR nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 15d04bfa-35cb-4734-9244-a4ecfe472fa3, please check neutron logs for more information. [ 573.990582] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Traceback (most recent call last): [ 573.990582] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.990582] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] self.driver.spawn(context, instance, image_meta, [ 573.990582] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 573.990582] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.990582] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.990582] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] vm_ref = self.build_virtual_machine(instance, [ 573.990582] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.990582] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.990582] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.991207] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] for vif in network_info: [ 573.991207] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.991207] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] return self._sync_wrapper(fn, *args, **kwargs) [ 573.991207] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.991207] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] self.wait() [ 573.991207] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.991207] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] self[:] = self._gt.wait() [ 573.991207] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.991207] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] return self._exit_event.wait() [ 573.991207] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.991207] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] result = hub.switch() [ 573.991207] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.991207] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] return self.greenlet.switch() [ 573.991580] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.991580] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] result = function(*args, **kwargs) [ 573.991580] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 573.991580] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] return func(*args, **kwargs) [ 573.991580] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.991580] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] raise e [ 573.991580] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.991580] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] nwinfo = self.network_api.allocate_for_instance( [ 573.991580] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.991580] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] created_port_ids = self._update_ports_for_instance( [ 573.991580] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.991580] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] with excutils.save_and_reraise_exception(): [ 573.991580] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.991989] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] self.force_reraise() [ 573.991989] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.991989] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] raise self.value [ 573.991989] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.991989] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] updated_port = self._update_port( [ 573.991989] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.991989] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] _ensure_no_port_binding_failure(port) [ 573.991989] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.991989] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] raise exception.PortBindingFailed(port_id=port['id']) [ 573.991989] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] nova.exception.PortBindingFailed: Binding failed for port 15d04bfa-35cb-4734-9244-a4ecfe472fa3, please check neutron logs for more information. [ 573.991989] env[61906]: ERROR nova.compute.manager [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] [ 573.992322] env[61906]: DEBUG nova.compute.utils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Binding failed for port 15d04bfa-35cb-4734-9244-a4ecfe472fa3, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 573.994491] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.305s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.995243] env[61906]: INFO nova.compute.claims [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 573.998908] env[61906]: DEBUG nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Build of instance 6c354d85-0252-4eb9-81e2-550a403bdbc5 was re-scheduled: Binding failed for port 15d04bfa-35cb-4734-9244-a4ecfe472fa3, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 573.999423] env[61906]: DEBUG nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 574.001040] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Acquiring lock "refresh_cache-6c354d85-0252-4eb9-81e2-550a403bdbc5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.001040] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Acquired lock "refresh_cache-6c354d85-0252-4eb9-81e2-550a403bdbc5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.001040] env[61906]: DEBUG nova.network.neutron [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.016394] env[61906]: DEBUG nova.network.neutron [-] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.019922] env[61906]: DEBUG nova.virt.hardware [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 574.020187] env[61906]: DEBUG nova.virt.hardware [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 574.020332] env[61906]: DEBUG nova.virt.hardware [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 574.020510] env[61906]: DEBUG nova.virt.hardware [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 574.020654] env[61906]: DEBUG nova.virt.hardware [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 574.020846] env[61906]: DEBUG nova.virt.hardware [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 574.020996] env[61906]: DEBUG nova.virt.hardware [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 574.021189] env[61906]: DEBUG nova.virt.hardware [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 574.021330] env[61906]: DEBUG nova.virt.hardware [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 574.021490] env[61906]: DEBUG nova.virt.hardware [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 574.021659] env[61906]: DEBUG nova.virt.hardware [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 574.022914] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb05cb31-7f32-44c9-981a-db5fe384233f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.032494] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6860a3a1-cd6c-4b3b-88f0-a0e8643c7dab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.383728] env[61906]: DEBUG nova.compute.manager [req-b6c80c1b-b031-4955-a3df-5d3863fa2170 req-dc58cf81-0de2-485a-b558-a3bf2497ce03 service nova] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Received event network-changed-c3930713-8fe6-470e-a2e4-6fe6c9377791 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.384088] env[61906]: DEBUG nova.compute.manager [req-b6c80c1b-b031-4955-a3df-5d3863fa2170 req-dc58cf81-0de2-485a-b558-a3bf2497ce03 service nova] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Refreshing instance network info cache due to event network-changed-c3930713-8fe6-470e-a2e4-6fe6c9377791. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 574.384355] env[61906]: DEBUG oslo_concurrency.lockutils [req-b6c80c1b-b031-4955-a3df-5d3863fa2170 req-dc58cf81-0de2-485a-b558-a3bf2497ce03 service nova] Acquiring lock "refresh_cache-6b3ecea4-fe20-4b6c-a987-59c20731b9ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.384704] env[61906]: DEBUG oslo_concurrency.lockutils [req-b6c80c1b-b031-4955-a3df-5d3863fa2170 req-dc58cf81-0de2-485a-b558-a3bf2497ce03 service nova] Acquired lock "refresh_cache-6b3ecea4-fe20-4b6c-a987-59c20731b9ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.384982] env[61906]: DEBUG nova.network.neutron [req-b6c80c1b-b031-4955-a3df-5d3863fa2170 req-dc58cf81-0de2-485a-b558-a3bf2497ce03 service nova] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Refreshing network info cache for port c3930713-8fe6-470e-a2e4-6fe6c9377791 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 574.528213] env[61906]: DEBUG nova.network.neutron [-] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.544371] env[61906]: DEBUG nova.network.neutron [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.829957] env[61906]: DEBUG nova.network.neutron [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.923207] env[61906]: DEBUG nova.network.neutron [req-b6c80c1b-b031-4955-a3df-5d3863fa2170 req-dc58cf81-0de2-485a-b558-a3bf2497ce03 service nova] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.031432] env[61906]: INFO nova.compute.manager [-] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Took 1.05 seconds to deallocate network for instance. [ 575.034228] env[61906]: DEBUG nova.compute.claims [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 575.034807] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.115761] env[61906]: DEBUG nova.network.neutron [req-b6c80c1b-b031-4955-a3df-5d3863fa2170 req-dc58cf81-0de2-485a-b558-a3bf2497ce03 service nova] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.210169] env[61906]: DEBUG nova.compute.manager [req-07914566-c6f2-4398-a828-1ca909d3adfb req-b2354356-caa0-4990-b483-7011eec97dd7 service nova] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Received event network-changed-208515e6-c9dd-4465-b0b9-0c97b49dab36 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 575.210169] env[61906]: DEBUG nova.compute.manager [req-07914566-c6f2-4398-a828-1ca909d3adfb req-b2354356-caa0-4990-b483-7011eec97dd7 service nova] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Refreshing instance network info cache due to event network-changed-208515e6-c9dd-4465-b0b9-0c97b49dab36. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 575.210169] env[61906]: DEBUG oslo_concurrency.lockutils [req-07914566-c6f2-4398-a828-1ca909d3adfb req-b2354356-caa0-4990-b483-7011eec97dd7 service nova] Acquiring lock "refresh_cache-d23d0d19-4806-4d5e-aa05-13aaf3c4a770" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.210419] env[61906]: DEBUG oslo_concurrency.lockutils [req-07914566-c6f2-4398-a828-1ca909d3adfb req-b2354356-caa0-4990-b483-7011eec97dd7 service nova] Acquired lock "refresh_cache-d23d0d19-4806-4d5e-aa05-13aaf3c4a770" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.210455] env[61906]: DEBUG nova.network.neutron [req-07914566-c6f2-4398-a828-1ca909d3adfb req-b2354356-caa0-4990-b483-7011eec97dd7 service nova] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Refreshing network info cache for port 208515e6-c9dd-4465-b0b9-0c97b49dab36 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 575.335846] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Releasing lock "refresh_cache-6c354d85-0252-4eb9-81e2-550a403bdbc5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.336141] env[61906]: DEBUG nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 575.336343] env[61906]: DEBUG nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 575.336511] env[61906]: DEBUG nova.network.neutron [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 575.355080] env[61906]: ERROR nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 208515e6-c9dd-4465-b0b9-0c97b49dab36, please check neutron logs for more information. [ 575.355080] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 575.355080] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.355080] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 575.355080] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.355080] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 575.355080] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.355080] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 575.355080] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.355080] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 575.355080] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.355080] env[61906]: ERROR nova.compute.manager raise self.value [ 575.355080] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.355080] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 575.355080] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.355080] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 575.355605] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.355605] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 575.355605] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 208515e6-c9dd-4465-b0b9-0c97b49dab36, please check neutron logs for more information. [ 575.355605] env[61906]: ERROR nova.compute.manager [ 575.355605] env[61906]: Traceback (most recent call last): [ 575.355605] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 575.355605] env[61906]: listener.cb(fileno) [ 575.355605] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.355605] env[61906]: result = function(*args, **kwargs) [ 575.355605] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 575.355605] env[61906]: return func(*args, **kwargs) [ 575.355605] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.355605] env[61906]: raise e [ 575.355605] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.355605] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 575.355605] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.355605] env[61906]: created_port_ids = self._update_ports_for_instance( [ 575.355605] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.355605] env[61906]: with excutils.save_and_reraise_exception(): [ 575.355605] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.355605] env[61906]: self.force_reraise() [ 575.355605] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.355605] env[61906]: raise self.value [ 575.355605] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.355605] env[61906]: updated_port = self._update_port( [ 575.355605] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.355605] env[61906]: _ensure_no_port_binding_failure(port) [ 575.355605] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.355605] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 575.394270] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 208515e6-c9dd-4465-b0b9-0c97b49dab36, please check neutron logs for more information. [ 575.394270] env[61906]: Removing descriptor: 15 [ 575.394270] env[61906]: ERROR nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 208515e6-c9dd-4465-b0b9-0c97b49dab36, please check neutron logs for more information. [ 575.394270] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Traceback (most recent call last): [ 575.394270] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 575.394270] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] yield resources [ 575.394270] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.394270] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] self.driver.spawn(context, instance, image_meta, [ 575.394270] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 575.394270] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.394270] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.394270] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] vm_ref = self.build_virtual_machine(instance, [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] for vif in network_info: [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] return self._sync_wrapper(fn, *args, **kwargs) [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] self.wait() [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] self[:] = self._gt.wait() [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] return self._exit_event.wait() [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.395676] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] result = hub.switch() [ 575.398469] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.398469] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] return self.greenlet.switch() [ 575.398469] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.398469] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] result = function(*args, **kwargs) [ 575.398469] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 575.398469] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] return func(*args, **kwargs) [ 575.398469] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.398469] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] raise e [ 575.398469] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.398469] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] nwinfo = self.network_api.allocate_for_instance( [ 575.398469] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.398469] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] created_port_ids = self._update_ports_for_instance( [ 575.398469] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.398835] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] with excutils.save_and_reraise_exception(): [ 575.398835] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.398835] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] self.force_reraise() [ 575.398835] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.398835] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] raise self.value [ 575.398835] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.398835] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] updated_port = self._update_port( [ 575.398835] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.398835] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] _ensure_no_port_binding_failure(port) [ 575.398835] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.398835] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] raise exception.PortBindingFailed(port_id=port['id']) [ 575.398835] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] nova.exception.PortBindingFailed: Binding failed for port 208515e6-c9dd-4465-b0b9-0c97b49dab36, please check neutron logs for more information. [ 575.398835] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] [ 575.399247] env[61906]: INFO nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Terminating instance [ 575.399247] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Acquiring lock "refresh_cache-d23d0d19-4806-4d5e-aa05-13aaf3c4a770" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.399247] env[61906]: DEBUG nova.network.neutron [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.495636] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06ddfc2a-1894-4143-91a7-a29dc0ebfd83 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.505269] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a6a2e7-35f5-4adb-a555-c41cb260133a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.539828] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16e7657-7aa9-4397-993e-aaf42afd0f31 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.555699] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b7d1e6-c1ba-4fcb-a86d-73da78aad052 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.568118] env[61906]: DEBUG nova.compute.provider_tree [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.619306] env[61906]: DEBUG oslo_concurrency.lockutils [req-b6c80c1b-b031-4955-a3df-5d3863fa2170 req-dc58cf81-0de2-485a-b558-a3bf2497ce03 service nova] Releasing lock "refresh_cache-6b3ecea4-fe20-4b6c-a987-59c20731b9ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.619306] env[61906]: DEBUG nova.compute.manager [req-b6c80c1b-b031-4955-a3df-5d3863fa2170 req-dc58cf81-0de2-485a-b558-a3bf2497ce03 service nova] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Received event network-vif-deleted-c3930713-8fe6-470e-a2e4-6fe6c9377791 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 575.744189] env[61906]: DEBUG nova.network.neutron [req-07914566-c6f2-4398-a828-1ca909d3adfb req-b2354356-caa0-4990-b483-7011eec97dd7 service nova] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.833493] env[61906]: DEBUG nova.network.neutron [req-07914566-c6f2-4398-a828-1ca909d3adfb req-b2354356-caa0-4990-b483-7011eec97dd7 service nova] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.871545] env[61906]: DEBUG nova.network.neutron [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.074644] env[61906]: DEBUG nova.scheduler.client.report [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.337814] env[61906]: DEBUG oslo_concurrency.lockutils [req-07914566-c6f2-4398-a828-1ca909d3adfb req-b2354356-caa0-4990-b483-7011eec97dd7 service nova] Releasing lock "refresh_cache-d23d0d19-4806-4d5e-aa05-13aaf3c4a770" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.338259] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Acquired lock "refresh_cache-d23d0d19-4806-4d5e-aa05-13aaf3c4a770" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.338445] env[61906]: DEBUG nova.network.neutron [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 576.373767] env[61906]: INFO nova.compute.manager [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] [instance: 6c354d85-0252-4eb9-81e2-550a403bdbc5] Took 1.04 seconds to deallocate network for instance. [ 576.579716] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.580257] env[61906]: DEBUG nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 576.582984] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.831s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.865583] env[61906]: DEBUG nova.network.neutron [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.985920] env[61906]: DEBUG nova.network.neutron [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.990091] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Acquiring lock "cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.990091] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Lock "cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.087812] env[61906]: DEBUG nova.compute.utils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 577.092148] env[61906]: DEBUG nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 577.092313] env[61906]: DEBUG nova.network.neutron [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 577.222125] env[61906]: DEBUG nova.policy [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a46c8a399120463284b3980744b6a5c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fbb116811b9f446484cb5439066cdc1c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 577.388666] env[61906]: DEBUG nova.compute.manager [req-b0b6b5df-6dae-40db-8a61-d4a94ecdeead req-47ef6582-10bb-4f79-82f4-cbd148d5eb69 service nova] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Received event network-vif-deleted-208515e6-c9dd-4465-b0b9-0c97b49dab36 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 577.415834] env[61906]: INFO nova.scheduler.client.report [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Deleted allocations for instance 6c354d85-0252-4eb9-81e2-550a403bdbc5 [ 577.491813] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Releasing lock "refresh_cache-d23d0d19-4806-4d5e-aa05-13aaf3c4a770" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.492280] env[61906]: DEBUG nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 577.492477] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 577.492766] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b25269a1-52e8-4b5d-b9c1-d0aed2d610d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.502976] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf518e70-97da-4146-a30f-7fb4c0346ff3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.515698] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d52ab0-8eac-42ea-9461-42c49b06e90f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.523875] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f090cee-428a-4bb1-94de-96f839e1e513 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.531240] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d23d0d19-4806-4d5e-aa05-13aaf3c4a770 could not be found. [ 577.531459] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 577.531636] env[61906]: INFO nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Took 0.04 seconds to destroy the instance on the hypervisor. [ 577.531873] env[61906]: DEBUG oslo.service.loopingcall [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 577.532504] env[61906]: DEBUG nova.compute.manager [-] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 577.532620] env[61906]: DEBUG nova.network.neutron [-] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 577.560655] env[61906]: DEBUG nova.network.neutron [-] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.562622] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a0d5ba-c74b-4037-93fa-a110dcdc2790 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.570858] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8bff70-ff26-4cd7-a7c6-06a355ad4546 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.584343] env[61906]: DEBUG nova.compute.provider_tree [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.595873] env[61906]: DEBUG nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 577.928599] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96744a8c-3778-4428-847b-a671d4800da6 tempest-AttachInterfacesUnderV243Test-56631782 tempest-AttachInterfacesUnderV243Test-56631782-project-member] Lock "6c354d85-0252-4eb9-81e2-550a403bdbc5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.219s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.066406] env[61906]: DEBUG nova.network.neutron [-] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.086676] env[61906]: DEBUG nova.scheduler.client.report [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 578.117562] env[61906]: DEBUG nova.network.neutron [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Successfully created port: a477cd11-7da0-45e8-8670-8c0022663392 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 578.433015] env[61906]: DEBUG nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 578.569896] env[61906]: INFO nova.compute.manager [-] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Took 1.04 seconds to deallocate network for instance. [ 578.572242] env[61906]: DEBUG nova.compute.claims [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 578.572438] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.591428] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.008s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.592146] env[61906]: ERROR nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3cee9f61-8537-4ca1-bef4-683dfe02020c, please check neutron logs for more information. [ 578.592146] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Traceback (most recent call last): [ 578.592146] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 578.592146] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] self.driver.spawn(context, instance, image_meta, [ 578.592146] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 578.592146] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] self._vmops.spawn(context, instance, image_meta, injected_files, [ 578.592146] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 578.592146] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] vm_ref = self.build_virtual_machine(instance, [ 578.592146] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 578.592146] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] vif_infos = vmwarevif.get_vif_info(self._session, [ 578.592146] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 578.593594] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] for vif in network_info: [ 578.593594] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 578.593594] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] return self._sync_wrapper(fn, *args, **kwargs) [ 578.593594] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 578.593594] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] self.wait() [ 578.593594] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 578.593594] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] self[:] = self._gt.wait() [ 578.593594] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 578.593594] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] return self._exit_event.wait() [ 578.593594] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 578.593594] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] result = hub.switch() [ 578.593594] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 578.593594] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] return self.greenlet.switch() [ 578.593883] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.593883] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] result = function(*args, **kwargs) [ 578.593883] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.593883] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] return func(*args, **kwargs) [ 578.593883] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.593883] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] raise e [ 578.593883] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.593883] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] nwinfo = self.network_api.allocate_for_instance( [ 578.593883] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.593883] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] created_port_ids = self._update_ports_for_instance( [ 578.593883] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.593883] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] with excutils.save_and_reraise_exception(): [ 578.593883] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.594153] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] self.force_reraise() [ 578.594153] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.594153] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] raise self.value [ 578.594153] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.594153] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] updated_port = self._update_port( [ 578.594153] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.594153] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] _ensure_no_port_binding_failure(port) [ 578.594153] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.594153] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] raise exception.PortBindingFailed(port_id=port['id']) [ 578.594153] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] nova.exception.PortBindingFailed: Binding failed for port 3cee9f61-8537-4ca1-bef4-683dfe02020c, please check neutron logs for more information. [ 578.594153] env[61906]: ERROR nova.compute.manager [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] [ 578.594456] env[61906]: DEBUG nova.compute.utils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Binding failed for port 3cee9f61-8537-4ca1-bef4-683dfe02020c, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 578.595596] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.990s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.595596] env[61906]: DEBUG nova.objects.instance [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Lazy-loading 'resources' on Instance uuid e23fd777-66f7-4ee1-bdcb-fe19c55f955c {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 578.596800] env[61906]: DEBUG nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Build of instance f708ec4a-7f1e-4bfa-b561-4b9bc5556d27 was re-scheduled: Binding failed for port 3cee9f61-8537-4ca1-bef4-683dfe02020c, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 578.597236] env[61906]: DEBUG nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 578.597460] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquiring lock "refresh_cache-f708ec4a-7f1e-4bfa-b561-4b9bc5556d27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.597604] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquired lock "refresh_cache-f708ec4a-7f1e-4bfa-b561-4b9bc5556d27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.597767] env[61906]: DEBUG nova.network.neutron [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 578.608212] env[61906]: DEBUG nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 578.647779] env[61906]: DEBUG nova.virt.hardware [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 578.648098] env[61906]: DEBUG nova.virt.hardware [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 578.648273] env[61906]: DEBUG nova.virt.hardware [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 578.648460] env[61906]: DEBUG nova.virt.hardware [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 578.648607] env[61906]: DEBUG nova.virt.hardware [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 578.648754] env[61906]: DEBUG nova.virt.hardware [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 578.648965] env[61906]: DEBUG nova.virt.hardware [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 578.649140] env[61906]: DEBUG nova.virt.hardware [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 578.649310] env[61906]: DEBUG nova.virt.hardware [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 578.649474] env[61906]: DEBUG nova.virt.hardware [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 578.650010] env[61906]: DEBUG nova.virt.hardware [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 578.651254] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396dfb82-176e-48a8-9a27-97a77101d779 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.661181] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74dd22cd-50e9-4cef-aa45-2b521c5b8481 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.965340] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.126968] env[61906]: DEBUG nova.network.neutron [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.268893] env[61906]: DEBUG nova.network.neutron [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.528272] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0f0033-d18a-4f05-89c7-bfc2b05e7d6e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.537366] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3478c40-8e2e-4b1e-9f57-e9c4d833c68d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.574299] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a53b439-5987-45af-a003-22328e010be8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.582339] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286384dd-f087-4d40-8c75-edec91d520d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.595667] env[61906]: DEBUG nova.compute.provider_tree [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.641902] env[61906]: DEBUG nova.compute.manager [req-de7ea748-c3c7-4ebe-b831-3c4f70d5dc3c req-388c48aa-7fe2-48cc-a080-87c41fc4b430 service nova] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Received event network-changed-a477cd11-7da0-45e8-8670-8c0022663392 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.642118] env[61906]: DEBUG nova.compute.manager [req-de7ea748-c3c7-4ebe-b831-3c4f70d5dc3c req-388c48aa-7fe2-48cc-a080-87c41fc4b430 service nova] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Refreshing instance network info cache due to event network-changed-a477cd11-7da0-45e8-8670-8c0022663392. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 579.642330] env[61906]: DEBUG oslo_concurrency.lockutils [req-de7ea748-c3c7-4ebe-b831-3c4f70d5dc3c req-388c48aa-7fe2-48cc-a080-87c41fc4b430 service nova] Acquiring lock "refresh_cache-528a2095-e82e-44eb-a4a1-dacde88966bb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.642793] env[61906]: DEBUG oslo_concurrency.lockutils [req-de7ea748-c3c7-4ebe-b831-3c4f70d5dc3c req-388c48aa-7fe2-48cc-a080-87c41fc4b430 service nova] Acquired lock "refresh_cache-528a2095-e82e-44eb-a4a1-dacde88966bb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.643155] env[61906]: DEBUG nova.network.neutron [req-de7ea748-c3c7-4ebe-b831-3c4f70d5dc3c req-388c48aa-7fe2-48cc-a080-87c41fc4b430 service nova] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Refreshing network info cache for port a477cd11-7da0-45e8-8670-8c0022663392 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 579.712556] env[61906]: ERROR nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a477cd11-7da0-45e8-8670-8c0022663392, please check neutron logs for more information. [ 579.712556] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 579.712556] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.712556] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 579.712556] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.712556] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 579.712556] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.712556] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 579.712556] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.712556] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 579.712556] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.712556] env[61906]: ERROR nova.compute.manager raise self.value [ 579.712556] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.712556] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 579.712556] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.712556] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 579.713198] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.713198] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 579.713198] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a477cd11-7da0-45e8-8670-8c0022663392, please check neutron logs for more information. [ 579.713198] env[61906]: ERROR nova.compute.manager [ 579.713198] env[61906]: Traceback (most recent call last): [ 579.713198] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 579.713198] env[61906]: listener.cb(fileno) [ 579.713198] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.713198] env[61906]: result = function(*args, **kwargs) [ 579.713198] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 579.713198] env[61906]: return func(*args, **kwargs) [ 579.713198] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.713198] env[61906]: raise e [ 579.713198] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.713198] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 579.713198] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.713198] env[61906]: created_port_ids = self._update_ports_for_instance( [ 579.713198] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.713198] env[61906]: with excutils.save_and_reraise_exception(): [ 579.713198] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.713198] env[61906]: self.force_reraise() [ 579.713198] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.713198] env[61906]: raise self.value [ 579.713198] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.713198] env[61906]: updated_port = self._update_port( [ 579.713198] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.713198] env[61906]: _ensure_no_port_binding_failure(port) [ 579.713198] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.713198] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 579.713855] env[61906]: nova.exception.PortBindingFailed: Binding failed for port a477cd11-7da0-45e8-8670-8c0022663392, please check neutron logs for more information. [ 579.713855] env[61906]: Removing descriptor: 17 [ 579.713855] env[61906]: ERROR nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a477cd11-7da0-45e8-8670-8c0022663392, please check neutron logs for more information. [ 579.713855] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Traceback (most recent call last): [ 579.713855] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 579.713855] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] yield resources [ 579.713855] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.713855] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] self.driver.spawn(context, instance, image_meta, [ 579.713855] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 579.713855] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.713855] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.713855] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] vm_ref = self.build_virtual_machine(instance, [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] for vif in network_info: [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] return self._sync_wrapper(fn, *args, **kwargs) [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] self.wait() [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] self[:] = self._gt.wait() [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] return self._exit_event.wait() [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.714120] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] result = hub.switch() [ 579.714391] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.714391] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] return self.greenlet.switch() [ 579.714391] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.714391] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] result = function(*args, **kwargs) [ 579.714391] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 579.714391] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] return func(*args, **kwargs) [ 579.714391] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.714391] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] raise e [ 579.714391] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.714391] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] nwinfo = self.network_api.allocate_for_instance( [ 579.714391] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.714391] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] created_port_ids = self._update_ports_for_instance( [ 579.714391] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.714646] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] with excutils.save_and_reraise_exception(): [ 579.714646] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.714646] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] self.force_reraise() [ 579.714646] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.714646] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] raise self.value [ 579.714646] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.714646] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] updated_port = self._update_port( [ 579.714646] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.714646] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] _ensure_no_port_binding_failure(port) [ 579.714646] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.714646] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] raise exception.PortBindingFailed(port_id=port['id']) [ 579.714646] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] nova.exception.PortBindingFailed: Binding failed for port a477cd11-7da0-45e8-8670-8c0022663392, please check neutron logs for more information. [ 579.714646] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] [ 579.714919] env[61906]: INFO nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Terminating instance [ 579.716063] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Acquiring lock "refresh_cache-528a2095-e82e-44eb-a4a1-dacde88966bb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.772140] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Releasing lock "refresh_cache-f708ec4a-7f1e-4bfa-b561-4b9bc5556d27" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.772417] env[61906]: DEBUG nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 579.772570] env[61906]: DEBUG nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 579.772732] env[61906]: DEBUG nova.network.neutron [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.800854] env[61906]: DEBUG nova.network.neutron [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.099892] env[61906]: DEBUG nova.scheduler.client.report [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.165862] env[61906]: DEBUG nova.network.neutron [req-de7ea748-c3c7-4ebe-b831-3c4f70d5dc3c req-388c48aa-7fe2-48cc-a080-87c41fc4b430 service nova] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.261360] env[61906]: DEBUG nova.network.neutron [req-de7ea748-c3c7-4ebe-b831-3c4f70d5dc3c req-388c48aa-7fe2-48cc-a080-87c41fc4b430 service nova] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.304317] env[61906]: DEBUG nova.network.neutron [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.609033] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.014s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.611314] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.131s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.616908] env[61906]: INFO nova.compute.claims [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 580.634212] env[61906]: INFO nova.scheduler.client.report [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Deleted allocations for instance e23fd777-66f7-4ee1-bdcb-fe19c55f955c [ 580.767790] env[61906]: DEBUG oslo_concurrency.lockutils [req-de7ea748-c3c7-4ebe-b831-3c4f70d5dc3c req-388c48aa-7fe2-48cc-a080-87c41fc4b430 service nova] Releasing lock "refresh_cache-528a2095-e82e-44eb-a4a1-dacde88966bb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.768261] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Acquired lock "refresh_cache-528a2095-e82e-44eb-a4a1-dacde88966bb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.768566] env[61906]: DEBUG nova.network.neutron [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 580.810288] env[61906]: INFO nova.compute.manager [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: f708ec4a-7f1e-4bfa-b561-4b9bc5556d27] Took 1.03 seconds to deallocate network for instance. [ 581.143318] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a9bd65ed-13e8-4208-8e3d-435e3ce06be7 tempest-ServersAaction247Test-1497126805 tempest-ServersAaction247Test-1497126805-project-member] Lock "e23fd777-66f7-4ee1-bdcb-fe19c55f955c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.275s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.299467] env[61906]: DEBUG nova.network.neutron [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.443772] env[61906]: DEBUG nova.network.neutron [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.695030] env[61906]: DEBUG nova.compute.manager [req-d2b09039-43e2-4624-9055-a004b8168f32 req-4ecb7ed0-f98f-4d2e-9760-6e68784c10bc service nova] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Received event network-vif-deleted-a477cd11-7da0-45e8-8670-8c0022663392 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 581.851188] env[61906]: INFO nova.scheduler.client.report [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Deleted allocations for instance f708ec4a-7f1e-4bfa-b561-4b9bc5556d27 [ 581.946180] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Releasing lock "refresh_cache-528a2095-e82e-44eb-a4a1-dacde88966bb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.946606] env[61906]: DEBUG nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 581.946825] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 581.947185] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ea9e0964-c818-4204-b78c-0b917a255ea4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.959092] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d914471-ea12-4c9c-a922-354276406986 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.984654] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 528a2095-e82e-44eb-a4a1-dacde88966bb could not be found. [ 581.984957] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 581.985200] env[61906]: INFO nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 581.985482] env[61906]: DEBUG oslo.service.loopingcall [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 581.988162] env[61906]: DEBUG nova.compute.manager [-] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.988349] env[61906]: DEBUG nova.network.neutron [-] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 582.012117] env[61906]: DEBUG nova.network.neutron [-] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.142219] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8065c5-7739-42bf-af99-20c883dadfae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.150589] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b3bcbe-bd40-4020-b7d8-12b9a1c6d693 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.188253] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f84ffa-3c8a-438e-8a08-1d2dba40ead5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.195755] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eaf5ba8-5ae2-491a-b61b-9d0e042934fb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.209719] env[61906]: DEBUG nova.compute.provider_tree [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.365751] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9dec20c7-4452-4e87-97e7-32587109c261 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Lock "f708ec4a-7f1e-4bfa-b561-4b9bc5556d27" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.637s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.514799] env[61906]: DEBUG nova.network.neutron [-] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.713773] env[61906]: DEBUG nova.scheduler.client.report [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.872534] env[61906]: DEBUG nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 583.020787] env[61906]: INFO nova.compute.manager [-] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Took 1.03 seconds to deallocate network for instance. [ 583.024675] env[61906]: DEBUG nova.compute.claims [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 583.024907] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.224257] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.613s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.224834] env[61906]: DEBUG nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 583.232720] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.947s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.410138] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.734508] env[61906]: DEBUG nova.compute.utils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 583.739660] env[61906]: DEBUG nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 583.739885] env[61906]: DEBUG nova.network.neutron [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 583.907459] env[61906]: DEBUG nova.policy [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ced8ec3ae40e4d44a2ad41f74784b172', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ae83ff4f61b47ec884177f50ac9d303', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 584.026713] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquiring lock "746ec539-5e13-4a9f-af73-26b175ad1e41" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.027336] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Lock "746ec539-5e13-4a9f-af73-26b175ad1e41" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.200056] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95ce83a-5c2d-44c6-8f70-f694110b5120 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.210736] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023e4d96-0921-4f18-a4a6-d00eb5570b64 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.246870] env[61906]: DEBUG nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 584.250024] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22023682-cf85-442b-85f6-77f1dc1be378 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.261018] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c21b7e5-f7db-460c-b226-bfb4c23eb12c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.280042] env[61906]: DEBUG nova.compute.provider_tree [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.736751] env[61906]: DEBUG nova.network.neutron [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Successfully created port: 509913ec-88cc-40e8-a970-b3a19be91b8b {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 584.783191] env[61906]: DEBUG nova.scheduler.client.report [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.873556] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "40dae609-e7a5-4af2-84d9-a53f93d95743" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.873967] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "40dae609-e7a5-4af2-84d9-a53f93d95743" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.261059] env[61906]: DEBUG nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 585.294577] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.065s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.297139] env[61906]: ERROR nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 009e2ac7-b58e-4c4b-ba55-9b8ccda347fe, please check neutron logs for more information. [ 585.297139] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Traceback (most recent call last): [ 585.297139] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 585.297139] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] self.driver.spawn(context, instance, image_meta, [ 585.297139] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 585.297139] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.297139] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.297139] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] vm_ref = self.build_virtual_machine(instance, [ 585.297139] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.297139] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.297139] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.297414] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] for vif in network_info: [ 585.297414] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.297414] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] return self._sync_wrapper(fn, *args, **kwargs) [ 585.297414] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.297414] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] self.wait() [ 585.297414] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.297414] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] self[:] = self._gt.wait() [ 585.297414] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.297414] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] return self._exit_event.wait() [ 585.297414] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.297414] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] result = hub.switch() [ 585.297414] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.297414] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] return self.greenlet.switch() [ 585.297795] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.297795] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] result = function(*args, **kwargs) [ 585.297795] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 585.297795] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] return func(*args, **kwargs) [ 585.297795] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.297795] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] raise e [ 585.297795] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.297795] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] nwinfo = self.network_api.allocate_for_instance( [ 585.297795] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.297795] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] created_port_ids = self._update_ports_for_instance( [ 585.297795] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.297795] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] with excutils.save_and_reraise_exception(): [ 585.297795] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.298242] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] self.force_reraise() [ 585.298242] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.298242] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] raise self.value [ 585.298242] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.298242] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] updated_port = self._update_port( [ 585.298242] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.298242] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] _ensure_no_port_binding_failure(port) [ 585.298242] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.298242] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] raise exception.PortBindingFailed(port_id=port['id']) [ 585.298242] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] nova.exception.PortBindingFailed: Binding failed for port 009e2ac7-b58e-4c4b-ba55-9b8ccda347fe, please check neutron logs for more information. [ 585.298242] env[61906]: ERROR nova.compute.manager [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] [ 585.298599] env[61906]: DEBUG nova.compute.utils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Binding failed for port 009e2ac7-b58e-4c4b-ba55-9b8ccda347fe, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 585.299026] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.322s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.303539] env[61906]: INFO nova.compute.claims [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.306466] env[61906]: DEBUG nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Build of instance ec81c4ec-07c6-4b1f-8555-69902cff4765 was re-scheduled: Binding failed for port 009e2ac7-b58e-4c4b-ba55-9b8ccda347fe, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 585.306974] env[61906]: DEBUG nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 585.307277] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquiring lock "refresh_cache-ec81c4ec-07c6-4b1f-8555-69902cff4765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.307427] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquired lock "refresh_cache-ec81c4ec-07c6-4b1f-8555-69902cff4765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.307581] env[61906]: DEBUG nova.network.neutron [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.314019] env[61906]: DEBUG nova.virt.hardware [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 585.314019] env[61906]: DEBUG nova.virt.hardware [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 585.314019] env[61906]: DEBUG nova.virt.hardware [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 585.314300] env[61906]: DEBUG nova.virt.hardware [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 585.314300] env[61906]: DEBUG nova.virt.hardware [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 585.314300] env[61906]: DEBUG nova.virt.hardware [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 585.314300] env[61906]: DEBUG nova.virt.hardware [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 585.314300] env[61906]: DEBUG nova.virt.hardware [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 585.314448] env[61906]: DEBUG nova.virt.hardware [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 585.314448] env[61906]: DEBUG nova.virt.hardware [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 585.314448] env[61906]: DEBUG nova.virt.hardware [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 585.314448] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c66ccae-5fd3-4954-9136-8d97bd981807 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.321741] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223ce23d-248e-4ffe-8873-8288df5d0bed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.837133] env[61906]: DEBUG nova.network.neutron [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.166420] env[61906]: DEBUG nova.network.neutron [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.673032] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Releasing lock "refresh_cache-ec81c4ec-07c6-4b1f-8555-69902cff4765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.673032] env[61906]: DEBUG nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 586.673032] env[61906]: DEBUG nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 586.673129] env[61906]: DEBUG nova.network.neutron [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 586.725587] env[61906]: DEBUG nova.network.neutron [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.824871] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af69d53-0a69-421e-9f46-e9061113db8e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.833975] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc89a043-7f8f-406b-83a8-6b0b061997ca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.873145] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb97997-a88e-4d47-a07f-66e9210e4e65 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.877148] env[61906]: DEBUG nova.compute.manager [req-c17c6199-bad7-4b8d-8e58-24fd35083937 req-9e8f8043-6239-4757-8440-3f9bae08af44 service nova] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Received event network-changed-509913ec-88cc-40e8-a970-b3a19be91b8b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.877148] env[61906]: DEBUG nova.compute.manager [req-c17c6199-bad7-4b8d-8e58-24fd35083937 req-9e8f8043-6239-4757-8440-3f9bae08af44 service nova] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Refreshing instance network info cache due to event network-changed-509913ec-88cc-40e8-a970-b3a19be91b8b. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 586.877148] env[61906]: DEBUG oslo_concurrency.lockutils [req-c17c6199-bad7-4b8d-8e58-24fd35083937 req-9e8f8043-6239-4757-8440-3f9bae08af44 service nova] Acquiring lock "refresh_cache-e64fd277-10ae-494d-918a-2f3faa4e44d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.877282] env[61906]: DEBUG oslo_concurrency.lockutils [req-c17c6199-bad7-4b8d-8e58-24fd35083937 req-9e8f8043-6239-4757-8440-3f9bae08af44 service nova] Acquired lock "refresh_cache-e64fd277-10ae-494d-918a-2f3faa4e44d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.877421] env[61906]: DEBUG nova.network.neutron [req-c17c6199-bad7-4b8d-8e58-24fd35083937 req-9e8f8043-6239-4757-8440-3f9bae08af44 service nova] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Refreshing network info cache for port 509913ec-88cc-40e8-a970-b3a19be91b8b {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 586.885769] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84578ff-9087-4f0e-8162-f4bfa89d08a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.904466] env[61906]: DEBUG nova.compute.provider_tree [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.230980] env[61906]: DEBUG nova.network.neutron [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.239350] env[61906]: ERROR nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 509913ec-88cc-40e8-a970-b3a19be91b8b, please check neutron logs for more information. [ 587.239350] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 587.239350] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.239350] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 587.239350] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.239350] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 587.239350] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.239350] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 587.239350] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.239350] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 587.239350] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.239350] env[61906]: ERROR nova.compute.manager raise self.value [ 587.239350] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.239350] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 587.239350] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.239350] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 587.239743] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.239743] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 587.239743] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 509913ec-88cc-40e8-a970-b3a19be91b8b, please check neutron logs for more information. [ 587.239743] env[61906]: ERROR nova.compute.manager [ 587.239743] env[61906]: Traceback (most recent call last): [ 587.239743] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 587.239743] env[61906]: listener.cb(fileno) [ 587.239743] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.239743] env[61906]: result = function(*args, **kwargs) [ 587.239743] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.239743] env[61906]: return func(*args, **kwargs) [ 587.239743] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.239743] env[61906]: raise e [ 587.239743] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.239743] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 587.239743] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.239743] env[61906]: created_port_ids = self._update_ports_for_instance( [ 587.239743] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.239743] env[61906]: with excutils.save_and_reraise_exception(): [ 587.239743] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.239743] env[61906]: self.force_reraise() [ 587.239743] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.239743] env[61906]: raise self.value [ 587.239743] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.239743] env[61906]: updated_port = self._update_port( [ 587.239743] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.239743] env[61906]: _ensure_no_port_binding_failure(port) [ 587.239743] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.239743] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 587.240414] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 509913ec-88cc-40e8-a970-b3a19be91b8b, please check neutron logs for more information. [ 587.240414] env[61906]: Removing descriptor: 17 [ 587.240414] env[61906]: ERROR nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 509913ec-88cc-40e8-a970-b3a19be91b8b, please check neutron logs for more information. [ 587.240414] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Traceback (most recent call last): [ 587.240414] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 587.240414] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] yield resources [ 587.240414] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.240414] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] self.driver.spawn(context, instance, image_meta, [ 587.240414] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 587.240414] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.240414] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.240414] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] vm_ref = self.build_virtual_machine(instance, [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] for vif in network_info: [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] return self._sync_wrapper(fn, *args, **kwargs) [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] self.wait() [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] self[:] = self._gt.wait() [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] return self._exit_event.wait() [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.240667] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] result = hub.switch() [ 587.240964] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.240964] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] return self.greenlet.switch() [ 587.240964] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.240964] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] result = function(*args, **kwargs) [ 587.240964] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.240964] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] return func(*args, **kwargs) [ 587.240964] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.240964] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] raise e [ 587.240964] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.240964] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] nwinfo = self.network_api.allocate_for_instance( [ 587.240964] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.240964] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] created_port_ids = self._update_ports_for_instance( [ 587.240964] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.241432] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] with excutils.save_and_reraise_exception(): [ 587.241432] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.241432] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] self.force_reraise() [ 587.241432] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.241432] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] raise self.value [ 587.241432] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.241432] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] updated_port = self._update_port( [ 587.241432] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.241432] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] _ensure_no_port_binding_failure(port) [ 587.241432] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.241432] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] raise exception.PortBindingFailed(port_id=port['id']) [ 587.241432] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] nova.exception.PortBindingFailed: Binding failed for port 509913ec-88cc-40e8-a970-b3a19be91b8b, please check neutron logs for more information. [ 587.241432] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] [ 587.241748] env[61906]: INFO nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Terminating instance [ 587.246544] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquiring lock "refresh_cache-e64fd277-10ae-494d-918a-2f3faa4e44d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.400594] env[61906]: DEBUG nova.network.neutron [req-c17c6199-bad7-4b8d-8e58-24fd35083937 req-9e8f8043-6239-4757-8440-3f9bae08af44 service nova] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.410891] env[61906]: DEBUG nova.scheduler.client.report [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.461661] env[61906]: DEBUG nova.network.neutron [req-c17c6199-bad7-4b8d-8e58-24fd35083937 req-9e8f8043-6239-4757-8440-3f9bae08af44 service nova] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.734419] env[61906]: INFO nova.compute.manager [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: ec81c4ec-07c6-4b1f-8555-69902cff4765] Took 1.06 seconds to deallocate network for instance. [ 587.918981] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.618s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.918981] env[61906]: DEBUG nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 587.921614] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.968s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.923148] env[61906]: INFO nova.compute.claims [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 587.963933] env[61906]: DEBUG oslo_concurrency.lockutils [req-c17c6199-bad7-4b8d-8e58-24fd35083937 req-9e8f8043-6239-4757-8440-3f9bae08af44 service nova] Releasing lock "refresh_cache-e64fd277-10ae-494d-918a-2f3faa4e44d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.964513] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquired lock "refresh_cache-e64fd277-10ae-494d-918a-2f3faa4e44d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.964599] env[61906]: DEBUG nova.network.neutron [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.427903] env[61906]: DEBUG nova.compute.utils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.431563] env[61906]: DEBUG nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 588.431737] env[61906]: DEBUG nova.network.neutron [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 588.472109] env[61906]: DEBUG nova.policy [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '97cefcc4373e4a168f780598f56d134b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e706eb33ab8349d9ada10b16282ec168', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 588.498256] env[61906]: DEBUG nova.network.neutron [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.576730] env[61906]: DEBUG nova.network.neutron [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.778555] env[61906]: INFO nova.scheduler.client.report [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Deleted allocations for instance ec81c4ec-07c6-4b1f-8555-69902cff4765 [ 588.932483] env[61906]: DEBUG nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 589.084290] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Releasing lock "refresh_cache-e64fd277-10ae-494d-918a-2f3faa4e44d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.084290] env[61906]: DEBUG nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.084290] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 589.086591] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca9ac0bc-a911-45d1-8345-40a82fe81abe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.100334] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05f155d0-5f91-4405-a263-00b755574851 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.115385] env[61906]: DEBUG nova.network.neutron [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Successfully created port: 9f9923e9-da89-45a9-9699-906de88dc362 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 589.129659] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e64fd277-10ae-494d-918a-2f3faa4e44d1 could not be found. [ 589.129659] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 589.129899] env[61906]: INFO nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Took 0.05 seconds to destroy the instance on the hypervisor. [ 589.130206] env[61906]: DEBUG oslo.service.loopingcall [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.130478] env[61906]: DEBUG nova.compute.manager [-] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.130478] env[61906]: DEBUG nova.network.neutron [-] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 589.164936] env[61906]: DEBUG nova.network.neutron [-] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.230393] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Acquiring lock "59227a6d-e699-43e3-8f40-e8767ffdc938" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.230621] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Lock "59227a6d-e699-43e3-8f40-e8767ffdc938" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.289706] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8061fdf3-6a2a-4759-ae3e-c26f5ae8abe6 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Lock "ec81c4ec-07c6-4b1f-8555-69902cff4765" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.278s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.376027] env[61906]: DEBUG nova.compute.manager [req-1c89ec8f-0567-4a22-8821-1d625a2c30f6 req-34263934-04d4-4a23-99ee-b3e3147beec9 service nova] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Received event network-vif-deleted-509913ec-88cc-40e8-a970-b3a19be91b8b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 589.394604] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41581501-9386-43a8-87cf-cb7b347b718e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.405758] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1dfdf7-2e9e-40f5-8a01-7b803a06beca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.447660] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-480ee6ab-7008-446d-b9a7-02f0e10a879d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.459024] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b52896-c369-4c44-846a-64742be99e2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.476628] env[61906]: DEBUG nova.compute.provider_tree [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.673839] env[61906]: DEBUG nova.network.neutron [-] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.792101] env[61906]: DEBUG nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 589.938807] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "2fdb9569-9e1a-44cd-9c2d-b493bc90ce81" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.938807] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "2fdb9569-9e1a-44cd-9c2d-b493bc90ce81" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.948430] env[61906]: DEBUG nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 589.979381] env[61906]: DEBUG nova.virt.hardware [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 589.979640] env[61906]: DEBUG nova.virt.hardware [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 589.979791] env[61906]: DEBUG nova.virt.hardware [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 589.980517] env[61906]: DEBUG nova.virt.hardware [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 589.980517] env[61906]: DEBUG nova.virt.hardware [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 589.980517] env[61906]: DEBUG nova.virt.hardware [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 589.980517] env[61906]: DEBUG nova.virt.hardware [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 589.980680] env[61906]: DEBUG nova.virt.hardware [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 589.980760] env[61906]: DEBUG nova.virt.hardware [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 589.980915] env[61906]: DEBUG nova.virt.hardware [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 589.981114] env[61906]: DEBUG nova.virt.hardware [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 589.982025] env[61906]: DEBUG nova.scheduler.client.report [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.985749] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d81170-625f-418d-820f-5efabc9767b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.997198] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03cd14d-b3f1-49a8-bbf6-c4ad1fa12bd9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.178713] env[61906]: INFO nova.compute.manager [-] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Took 1.05 seconds to deallocate network for instance. [ 590.181891] env[61906]: DEBUG nova.compute.claims [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.182123] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.252265] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Acquiring lock "3c0602a1-ad7d-4f85-b4e0-d36565e28a38" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.252911] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Lock "3c0602a1-ad7d-4f85-b4e0-d36565e28a38" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.325128] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.389092] env[61906]: ERROR nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9f9923e9-da89-45a9-9699-906de88dc362, please check neutron logs for more information. [ 590.389092] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 590.389092] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.389092] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 590.389092] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 590.389092] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 590.389092] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 590.389092] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 590.389092] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.389092] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 590.389092] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.389092] env[61906]: ERROR nova.compute.manager raise self.value [ 590.389092] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 590.389092] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 590.389092] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.389092] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 590.389456] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.389456] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 590.389456] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9f9923e9-da89-45a9-9699-906de88dc362, please check neutron logs for more information. [ 590.389456] env[61906]: ERROR nova.compute.manager [ 590.389456] env[61906]: Traceback (most recent call last): [ 590.389456] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 590.389456] env[61906]: listener.cb(fileno) [ 590.389456] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.389456] env[61906]: result = function(*args, **kwargs) [ 590.389456] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 590.389456] env[61906]: return func(*args, **kwargs) [ 590.389456] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.389456] env[61906]: raise e [ 590.389456] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.389456] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 590.389456] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 590.389456] env[61906]: created_port_ids = self._update_ports_for_instance( [ 590.389456] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 590.389456] env[61906]: with excutils.save_and_reraise_exception(): [ 590.389456] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.389456] env[61906]: self.force_reraise() [ 590.389456] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.389456] env[61906]: raise self.value [ 590.389456] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 590.389456] env[61906]: updated_port = self._update_port( [ 590.389456] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.389456] env[61906]: _ensure_no_port_binding_failure(port) [ 590.389456] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.389456] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 590.390055] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 9f9923e9-da89-45a9-9699-906de88dc362, please check neutron logs for more information. [ 590.390055] env[61906]: Removing descriptor: 17 [ 590.390055] env[61906]: ERROR nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9f9923e9-da89-45a9-9699-906de88dc362, please check neutron logs for more information. [ 590.390055] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Traceback (most recent call last): [ 590.390055] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 590.390055] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] yield resources [ 590.390055] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 590.390055] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] self.driver.spawn(context, instance, image_meta, [ 590.390055] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 590.390055] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] self._vmops.spawn(context, instance, image_meta, injected_files, [ 590.390055] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 590.390055] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] vm_ref = self.build_virtual_machine(instance, [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] vif_infos = vmwarevif.get_vif_info(self._session, [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] for vif in network_info: [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] return self._sync_wrapper(fn, *args, **kwargs) [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] self.wait() [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] self[:] = self._gt.wait() [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] return self._exit_event.wait() [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 590.390312] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] result = hub.switch() [ 590.390611] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 590.390611] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] return self.greenlet.switch() [ 590.390611] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.390611] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] result = function(*args, **kwargs) [ 590.390611] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 590.390611] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] return func(*args, **kwargs) [ 590.390611] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.390611] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] raise e [ 590.390611] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.390611] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] nwinfo = self.network_api.allocate_for_instance( [ 590.390611] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 590.390611] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] created_port_ids = self._update_ports_for_instance( [ 590.390611] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 590.390884] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] with excutils.save_and_reraise_exception(): [ 590.390884] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.390884] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] self.force_reraise() [ 590.390884] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.390884] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] raise self.value [ 590.390884] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 590.390884] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] updated_port = self._update_port( [ 590.390884] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.390884] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] _ensure_no_port_binding_failure(port) [ 590.390884] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.390884] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] raise exception.PortBindingFailed(port_id=port['id']) [ 590.390884] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] nova.exception.PortBindingFailed: Binding failed for port 9f9923e9-da89-45a9-9699-906de88dc362, please check neutron logs for more information. [ 590.390884] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] [ 590.391169] env[61906]: INFO nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Terminating instance [ 590.392508] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Acquiring lock "refresh_cache-4b0abc14-0ce2-4212-8677-82f6279ee107" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.392717] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Acquired lock "refresh_cache-4b0abc14-0ce2-4212-8677-82f6279ee107" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.392909] env[61906]: DEBUG nova.network.neutron [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 590.493278] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.572s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.493871] env[61906]: DEBUG nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 590.497909] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.914s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.136876] env[61906]: DEBUG nova.compute.utils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.143049] env[61906]: DEBUG nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 591.143049] env[61906]: DEBUG nova.network.neutron [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 591.144434] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquiring lock "4a2ed7bd-eb49-49a0-ba81-3c29baea6533" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.144700] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Lock "4a2ed7bd-eb49-49a0-ba81-3c29baea6533" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.169013] env[61906]: DEBUG nova.network.neutron [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.212579] env[61906]: DEBUG nova.policy [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'af7de094ae154a8982934396eb566eb9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b453eac54a7047b696bda7f88c0c917e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 591.272841] env[61906]: DEBUG nova.network.neutron [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.302344] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquiring lock "0dca3209-d37b-4536-bbf9-72f0418cef8e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.302581] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Lock "0dca3209-d37b-4536-bbf9-72f0418cef8e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.399474] env[61906]: DEBUG nova.compute.manager [req-10674ebd-d3fa-4663-ba8b-60142c3295d7 req-f19a8001-5f21-4784-acdb-998fe3c5834f service nova] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Received event network-changed-9f9923e9-da89-45a9-9699-906de88dc362 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 591.399734] env[61906]: DEBUG nova.compute.manager [req-10674ebd-d3fa-4663-ba8b-60142c3295d7 req-f19a8001-5f21-4784-acdb-998fe3c5834f service nova] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Refreshing instance network info cache due to event network-changed-9f9923e9-da89-45a9-9699-906de88dc362. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 591.399939] env[61906]: DEBUG oslo_concurrency.lockutils [req-10674ebd-d3fa-4663-ba8b-60142c3295d7 req-f19a8001-5f21-4784-acdb-998fe3c5834f service nova] Acquiring lock "refresh_cache-4b0abc14-0ce2-4212-8677-82f6279ee107" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.525851] env[61906]: DEBUG nova.network.neutron [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Successfully created port: 49eae8f7-1b56-4377-9e75-390f8c019ca6 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 591.648466] env[61906]: DEBUG nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 591.778017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Releasing lock "refresh_cache-4b0abc14-0ce2-4212-8677-82f6279ee107" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.778017] env[61906]: DEBUG nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 591.778017] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 591.778017] env[61906]: DEBUG oslo_concurrency.lockutils [req-10674ebd-d3fa-4663-ba8b-60142c3295d7 req-f19a8001-5f21-4784-acdb-998fe3c5834f service nova] Acquired lock "refresh_cache-4b0abc14-0ce2-4212-8677-82f6279ee107" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.778017] env[61906]: DEBUG nova.network.neutron [req-10674ebd-d3fa-4663-ba8b-60142c3295d7 req-f19a8001-5f21-4784-acdb-998fe3c5834f service nova] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Refreshing network info cache for port 9f9923e9-da89-45a9-9699-906de88dc362 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 591.778249] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e99982e8-1b87-4d4d-bf2c-7a082954cddd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.793209] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e53f771d-a600-4c4d-afe0-51415ea45848 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.818829] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4b0abc14-0ce2-4212-8677-82f6279ee107 could not be found. [ 591.819051] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 591.819254] env[61906]: INFO nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Took 0.04 seconds to destroy the instance on the hypervisor. [ 591.819498] env[61906]: DEBUG oslo.service.loopingcall [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 591.822034] env[61906]: DEBUG nova.compute.manager [-] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 591.822172] env[61906]: DEBUG nova.network.neutron [-] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 591.836856] env[61906]: DEBUG nova.network.neutron [-] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.139266] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568d4ab7-eca6-444d-a605-8c2fce9ea01d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.147640] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee88a206-32af-4190-97a4-62526838a370 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.181099] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-511c6c20-7917-4684-84e0-0200e7914d60 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.188738] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a49fe6-781f-4364-b2a8-bb8402530230 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.202677] env[61906]: DEBUG nova.compute.provider_tree [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.302840] env[61906]: DEBUG nova.network.neutron [req-10674ebd-d3fa-4663-ba8b-60142c3295d7 req-f19a8001-5f21-4784-acdb-998fe3c5834f service nova] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.342123] env[61906]: DEBUG nova.network.neutron [-] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.416134] env[61906]: DEBUG nova.network.neutron [req-10674ebd-d3fa-4663-ba8b-60142c3295d7 req-f19a8001-5f21-4784-acdb-998fe3c5834f service nova] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.545384] env[61906]: DEBUG nova.compute.manager [req-d149ebf8-fa7f-4f38-b1a6-e25ab0e6e237 req-1f526a4c-49ba-4c4d-82c1-74cb1663b469 service nova] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Received event network-changed-49eae8f7-1b56-4377-9e75-390f8c019ca6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.545384] env[61906]: DEBUG nova.compute.manager [req-d149ebf8-fa7f-4f38-b1a6-e25ab0e6e237 req-1f526a4c-49ba-4c4d-82c1-74cb1663b469 service nova] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Refreshing instance network info cache due to event network-changed-49eae8f7-1b56-4377-9e75-390f8c019ca6. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 592.545384] env[61906]: DEBUG oslo_concurrency.lockutils [req-d149ebf8-fa7f-4f38-b1a6-e25ab0e6e237 req-1f526a4c-49ba-4c4d-82c1-74cb1663b469 service nova] Acquiring lock "refresh_cache-05cefc82-3585-4d0b-8620-91412b3ef69e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.545384] env[61906]: DEBUG oslo_concurrency.lockutils [req-d149ebf8-fa7f-4f38-b1a6-e25ab0e6e237 req-1f526a4c-49ba-4c4d-82c1-74cb1663b469 service nova] Acquired lock "refresh_cache-05cefc82-3585-4d0b-8620-91412b3ef69e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.545384] env[61906]: DEBUG nova.network.neutron [req-d149ebf8-fa7f-4f38-b1a6-e25ab0e6e237 req-1f526a4c-49ba-4c4d-82c1-74cb1663b469 service nova] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Refreshing network info cache for port 49eae8f7-1b56-4377-9e75-390f8c019ca6 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 592.682078] env[61906]: DEBUG nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 592.706041] env[61906]: DEBUG nova.scheduler.client.report [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.714020] env[61906]: DEBUG nova.virt.hardware [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 592.714020] env[61906]: DEBUG nova.virt.hardware [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 592.714020] env[61906]: DEBUG nova.virt.hardware [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 592.714248] env[61906]: DEBUG nova.virt.hardware [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 592.714248] env[61906]: DEBUG nova.virt.hardware [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 592.714248] env[61906]: DEBUG nova.virt.hardware [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 592.714325] env[61906]: DEBUG nova.virt.hardware [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 592.714445] env[61906]: DEBUG nova.virt.hardware [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 592.714572] env[61906]: DEBUG nova.virt.hardware [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 592.714724] env[61906]: DEBUG nova.virt.hardware [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 592.714886] env[61906]: DEBUG nova.virt.hardware [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 592.715761] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc145c7-e999-46fa-a039-40842ecf2338 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.723913] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493d2aeb-e5ac-428a-94d0-ffd0366d5156 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.743054] env[61906]: ERROR nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 49eae8f7-1b56-4377-9e75-390f8c019ca6, please check neutron logs for more information. [ 592.743054] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 592.743054] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.743054] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 592.743054] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.743054] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 592.743054] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.743054] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 592.743054] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.743054] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 592.743054] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.743054] env[61906]: ERROR nova.compute.manager raise self.value [ 592.743054] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.743054] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 592.743054] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.743054] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 592.743710] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.743710] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 592.743710] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 49eae8f7-1b56-4377-9e75-390f8c019ca6, please check neutron logs for more information. [ 592.743710] env[61906]: ERROR nova.compute.manager [ 592.743710] env[61906]: Traceback (most recent call last): [ 592.743710] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 592.743710] env[61906]: listener.cb(fileno) [ 592.743710] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.743710] env[61906]: result = function(*args, **kwargs) [ 592.743710] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.743710] env[61906]: return func(*args, **kwargs) [ 592.743710] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.743710] env[61906]: raise e [ 592.743710] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.743710] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 592.743710] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.743710] env[61906]: created_port_ids = self._update_ports_for_instance( [ 592.743710] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.743710] env[61906]: with excutils.save_and_reraise_exception(): [ 592.743710] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.743710] env[61906]: self.force_reraise() [ 592.743710] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.743710] env[61906]: raise self.value [ 592.743710] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.743710] env[61906]: updated_port = self._update_port( [ 592.743710] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.743710] env[61906]: _ensure_no_port_binding_failure(port) [ 592.743710] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.743710] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 592.744785] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 49eae8f7-1b56-4377-9e75-390f8c019ca6, please check neutron logs for more information. [ 592.744785] env[61906]: Removing descriptor: 15 [ 592.744785] env[61906]: ERROR nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 49eae8f7-1b56-4377-9e75-390f8c019ca6, please check neutron logs for more information. [ 592.744785] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Traceback (most recent call last): [ 592.744785] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 592.744785] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] yield resources [ 592.744785] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.744785] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] self.driver.spawn(context, instance, image_meta, [ 592.744785] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 592.744785] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.744785] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.744785] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] vm_ref = self.build_virtual_machine(instance, [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] for vif in network_info: [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] return self._sync_wrapper(fn, *args, **kwargs) [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] self.wait() [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] self[:] = self._gt.wait() [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] return self._exit_event.wait() [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.745251] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] result = hub.switch() [ 592.745810] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.745810] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] return self.greenlet.switch() [ 592.745810] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.745810] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] result = function(*args, **kwargs) [ 592.745810] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.745810] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] return func(*args, **kwargs) [ 592.745810] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.745810] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] raise e [ 592.745810] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.745810] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] nwinfo = self.network_api.allocate_for_instance( [ 592.745810] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.745810] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] created_port_ids = self._update_ports_for_instance( [ 592.745810] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.746228] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] with excutils.save_and_reraise_exception(): [ 592.746228] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.746228] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] self.force_reraise() [ 592.746228] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.746228] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] raise self.value [ 592.746228] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.746228] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] updated_port = self._update_port( [ 592.746228] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.746228] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] _ensure_no_port_binding_failure(port) [ 592.746228] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.746228] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] raise exception.PortBindingFailed(port_id=port['id']) [ 592.746228] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] nova.exception.PortBindingFailed: Binding failed for port 49eae8f7-1b56-4377-9e75-390f8c019ca6, please check neutron logs for more information. [ 592.746228] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] [ 592.746515] env[61906]: INFO nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Terminating instance [ 592.746515] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Acquiring lock "refresh_cache-05cefc82-3585-4d0b-8620-91412b3ef69e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.844941] env[61906]: INFO nova.compute.manager [-] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Took 1.02 seconds to deallocate network for instance. [ 592.847280] env[61906]: DEBUG nova.compute.claims [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 592.847453] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.920722] env[61906]: DEBUG oslo_concurrency.lockutils [req-10674ebd-d3fa-4663-ba8b-60142c3295d7 req-f19a8001-5f21-4784-acdb-998fe3c5834f service nova] Releasing lock "refresh_cache-4b0abc14-0ce2-4212-8677-82f6279ee107" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.921017] env[61906]: DEBUG nova.compute.manager [req-10674ebd-d3fa-4663-ba8b-60142c3295d7 req-f19a8001-5f21-4784-acdb-998fe3c5834f service nova] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Received event network-vif-deleted-9f9923e9-da89-45a9-9699-906de88dc362 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 593.065742] env[61906]: DEBUG nova.network.neutron [req-d149ebf8-fa7f-4f38-b1a6-e25ab0e6e237 req-1f526a4c-49ba-4c4d-82c1-74cb1663b469 service nova] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.143520] env[61906]: DEBUG nova.network.neutron [req-d149ebf8-fa7f-4f38-b1a6-e25ab0e6e237 req-1f526a4c-49ba-4c4d-82c1-74cb1663b469 service nova] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.220095] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.722s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.220752] env[61906]: ERROR nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port faa99eaa-4906-4750-b710-7771eb7cf572, please check neutron logs for more information. [ 593.220752] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Traceback (most recent call last): [ 593.220752] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.220752] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] self.driver.spawn(context, instance, image_meta, [ 593.220752] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 593.220752] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.220752] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.220752] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] vm_ref = self.build_virtual_machine(instance, [ 593.220752] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.220752] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.220752] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.221031] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] for vif in network_info: [ 593.221031] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 593.221031] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] return self._sync_wrapper(fn, *args, **kwargs) [ 593.221031] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 593.221031] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] self.wait() [ 593.221031] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 593.221031] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] self[:] = self._gt.wait() [ 593.221031] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.221031] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] return self._exit_event.wait() [ 593.221031] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.221031] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] result = hub.switch() [ 593.221031] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.221031] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] return self.greenlet.switch() [ 593.221346] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.221346] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] result = function(*args, **kwargs) [ 593.221346] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 593.221346] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] return func(*args, **kwargs) [ 593.221346] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.221346] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] raise e [ 593.221346] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.221346] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] nwinfo = self.network_api.allocate_for_instance( [ 593.221346] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.221346] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] created_port_ids = self._update_ports_for_instance( [ 593.221346] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.221346] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] with excutils.save_and_reraise_exception(): [ 593.221346] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.221683] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] self.force_reraise() [ 593.221683] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.221683] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] raise self.value [ 593.221683] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.221683] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] updated_port = self._update_port( [ 593.221683] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.221683] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] _ensure_no_port_binding_failure(port) [ 593.221683] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.221683] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] raise exception.PortBindingFailed(port_id=port['id']) [ 593.221683] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] nova.exception.PortBindingFailed: Binding failed for port faa99eaa-4906-4750-b710-7771eb7cf572, please check neutron logs for more information. [ 593.221683] env[61906]: ERROR nova.compute.manager [instance: f72e6538-7130-45a9-a722-ba60de3bda49] [ 593.221965] env[61906]: DEBUG nova.compute.utils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Binding failed for port faa99eaa-4906-4750-b710-7771eb7cf572, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 593.222790] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.188s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.225752] env[61906]: DEBUG nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Build of instance f72e6538-7130-45a9-a722-ba60de3bda49 was re-scheduled: Binding failed for port faa99eaa-4906-4750-b710-7771eb7cf572, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 593.226233] env[61906]: DEBUG nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 593.226442] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquiring lock "refresh_cache-f72e6538-7130-45a9-a722-ba60de3bda49" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.226583] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Acquired lock "refresh_cache-f72e6538-7130-45a9-a722-ba60de3bda49" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.226738] env[61906]: DEBUG nova.network.neutron [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 593.646191] env[61906]: DEBUG oslo_concurrency.lockutils [req-d149ebf8-fa7f-4f38-b1a6-e25ab0e6e237 req-1f526a4c-49ba-4c4d-82c1-74cb1663b469 service nova] Releasing lock "refresh_cache-05cefc82-3585-4d0b-8620-91412b3ef69e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.646596] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Acquired lock "refresh_cache-05cefc82-3585-4d0b-8620-91412b3ef69e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.646783] env[61906]: DEBUG nova.network.neutron [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 593.746889] env[61906]: DEBUG nova.network.neutron [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.831691] env[61906]: DEBUG nova.network.neutron [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.090537] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce20a7ba-7872-4076-a802-6ddb89d4ea81 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.098216] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc06492-bd38-448a-8c65-424d586dcfb7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.128952] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85a8bd3-c8af-40d9-ae40-c67234059893 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.135919] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a7d458b-1486-4cfa-b5ed-66e8999c97bb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.150882] env[61906]: DEBUG nova.compute.provider_tree [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.165989] env[61906]: DEBUG nova.network.neutron [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.259932] env[61906]: DEBUG nova.network.neutron [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.334282] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Releasing lock "refresh_cache-f72e6538-7130-45a9-a722-ba60de3bda49" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.334600] env[61906]: DEBUG nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 594.334802] env[61906]: DEBUG nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.334966] env[61906]: DEBUG nova.network.neutron [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 594.349819] env[61906]: DEBUG nova.network.neutron [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.569853] env[61906]: DEBUG nova.compute.manager [req-d8b62110-be9e-469b-98df-a927ac569b93 req-5be9a04b-17b6-4639-9f44-978a21534a80 service nova] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Received event network-vif-deleted-49eae8f7-1b56-4377-9e75-390f8c019ca6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 594.654469] env[61906]: DEBUG nova.scheduler.client.report [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.763172] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Releasing lock "refresh_cache-05cefc82-3585-4d0b-8620-91412b3ef69e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.763746] env[61906]: DEBUG nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 594.764014] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 594.764326] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17206507-cf9c-46be-bb17-4a239d98df72 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.773589] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6556be0-1f2d-4042-ab59-ebf5354bcb31 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.794638] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 05cefc82-3585-4d0b-8620-91412b3ef69e could not be found. [ 594.794853] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 594.795043] env[61906]: INFO nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 594.795293] env[61906]: DEBUG oslo.service.loopingcall [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 594.795505] env[61906]: DEBUG nova.compute.manager [-] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.795599] env[61906]: DEBUG nova.network.neutron [-] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 594.811690] env[61906]: DEBUG nova.network.neutron [-] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.851950] env[61906]: DEBUG nova.network.neutron [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.160471] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.938s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.161112] env[61906]: ERROR nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c3930713-8fe6-470e-a2e4-6fe6c9377791, please check neutron logs for more information. [ 595.161112] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Traceback (most recent call last): [ 595.161112] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 595.161112] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] self.driver.spawn(context, instance, image_meta, [ 595.161112] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 595.161112] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.161112] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.161112] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] vm_ref = self.build_virtual_machine(instance, [ 595.161112] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.161112] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.161112] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.161435] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] for vif in network_info: [ 595.161435] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.161435] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] return self._sync_wrapper(fn, *args, **kwargs) [ 595.161435] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.161435] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] self.wait() [ 595.161435] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.161435] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] self[:] = self._gt.wait() [ 595.161435] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.161435] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] return self._exit_event.wait() [ 595.161435] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.161435] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] result = hub.switch() [ 595.161435] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.161435] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] return self.greenlet.switch() [ 595.161769] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.161769] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] result = function(*args, **kwargs) [ 595.161769] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 595.161769] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] return func(*args, **kwargs) [ 595.161769] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.161769] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] raise e [ 595.161769] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.161769] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] nwinfo = self.network_api.allocate_for_instance( [ 595.161769] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.161769] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] created_port_ids = self._update_ports_for_instance( [ 595.161769] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.161769] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] with excutils.save_and_reraise_exception(): [ 595.161769] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.162135] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] self.force_reraise() [ 595.162135] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.162135] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] raise self.value [ 595.162135] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.162135] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] updated_port = self._update_port( [ 595.162135] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.162135] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] _ensure_no_port_binding_failure(port) [ 595.162135] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.162135] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] raise exception.PortBindingFailed(port_id=port['id']) [ 595.162135] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] nova.exception.PortBindingFailed: Binding failed for port c3930713-8fe6-470e-a2e4-6fe6c9377791, please check neutron logs for more information. [ 595.162135] env[61906]: ERROR nova.compute.manager [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] [ 595.162426] env[61906]: DEBUG nova.compute.utils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Binding failed for port c3930713-8fe6-470e-a2e4-6fe6c9377791, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 595.163583] env[61906]: DEBUG nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Build of instance 6b3ecea4-fe20-4b6c-a987-59c20731b9ac was re-scheduled: Binding failed for port c3930713-8fe6-470e-a2e4-6fe6c9377791, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 595.164119] env[61906]: DEBUG nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 595.164355] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Acquiring lock "refresh_cache-6b3ecea4-fe20-4b6c-a987-59c20731b9ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.164502] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Acquired lock "refresh_cache-6b3ecea4-fe20-4b6c-a987-59c20731b9ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.164659] env[61906]: DEBUG nova.network.neutron [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 595.166572] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.594s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.314435] env[61906]: DEBUG nova.network.neutron [-] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.355258] env[61906]: INFO nova.compute.manager [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] [instance: f72e6538-7130-45a9-a722-ba60de3bda49] Took 1.02 seconds to deallocate network for instance. [ 595.688817] env[61906]: DEBUG nova.network.neutron [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.770607] env[61906]: DEBUG nova.network.neutron [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.817487] env[61906]: INFO nova.compute.manager [-] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Took 1.02 seconds to deallocate network for instance. [ 595.819507] env[61906]: DEBUG nova.compute.claims [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 595.819689] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.022763] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb24c6c-4dab-4552-91d8-100eb098c04d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.031129] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ffee7b-6922-44f9-9de2-d6ac7bb0af1a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.059955] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fca076d-7662-4cc5-9cf8-d1d6bcbe4fe6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.067445] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f16eac-91b4-41af-b972-de714a463065 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.081378] env[61906]: DEBUG nova.compute.provider_tree [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.275553] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Releasing lock "refresh_cache-6b3ecea4-fe20-4b6c-a987-59c20731b9ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.276659] env[61906]: DEBUG nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 596.276659] env[61906]: DEBUG nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.276659] env[61906]: DEBUG nova.network.neutron [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 596.291161] env[61906]: DEBUG nova.network.neutron [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.383035] env[61906]: INFO nova.scheduler.client.report [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Deleted allocations for instance f72e6538-7130-45a9-a722-ba60de3bda49 [ 596.584382] env[61906]: DEBUG nova.scheduler.client.report [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.795802] env[61906]: DEBUG nova.network.neutron [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.893297] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa0ab248-d5e0-48b1-be41-9cd1cb310014 tempest-ServersAdminTestJSON-1498472260 tempest-ServersAdminTestJSON-1498472260-project-member] Lock "f72e6538-7130-45a9-a722-ba60de3bda49" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.204s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.090237] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.923s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.090237] env[61906]: ERROR nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 208515e6-c9dd-4465-b0b9-0c97b49dab36, please check neutron logs for more information. [ 597.090237] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Traceback (most recent call last): [ 597.090237] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.090237] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] self.driver.spawn(context, instance, image_meta, [ 597.090237] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 597.090237] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.090237] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.090237] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] vm_ref = self.build_virtual_machine(instance, [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] for vif in network_info: [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] return self._sync_wrapper(fn, *args, **kwargs) [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] self.wait() [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] self[:] = self._gt.wait() [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] return self._exit_event.wait() [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.090623] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] result = hub.switch() [ 597.090976] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.090976] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] return self.greenlet.switch() [ 597.090976] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.090976] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] result = function(*args, **kwargs) [ 597.090976] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 597.090976] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] return func(*args, **kwargs) [ 597.090976] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.090976] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] raise e [ 597.090976] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.090976] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] nwinfo = self.network_api.allocate_for_instance( [ 597.090976] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.090976] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] created_port_ids = self._update_ports_for_instance( [ 597.090976] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.091328] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] with excutils.save_and_reraise_exception(): [ 597.091328] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.091328] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] self.force_reraise() [ 597.091328] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.091328] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] raise self.value [ 597.091328] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.091328] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] updated_port = self._update_port( [ 597.091328] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.091328] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] _ensure_no_port_binding_failure(port) [ 597.091328] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.091328] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] raise exception.PortBindingFailed(port_id=port['id']) [ 597.091328] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] nova.exception.PortBindingFailed: Binding failed for port 208515e6-c9dd-4465-b0b9-0c97b49dab36, please check neutron logs for more information. [ 597.091328] env[61906]: ERROR nova.compute.manager [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] [ 597.091635] env[61906]: DEBUG nova.compute.utils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Binding failed for port 208515e6-c9dd-4465-b0b9-0c97b49dab36, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 597.091635] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.126s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.093086] env[61906]: INFO nova.compute.claims [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.096040] env[61906]: DEBUG nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Build of instance d23d0d19-4806-4d5e-aa05-13aaf3c4a770 was re-scheduled: Binding failed for port 208515e6-c9dd-4465-b0b9-0c97b49dab36, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 597.096305] env[61906]: DEBUG nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 597.097029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Acquiring lock "refresh_cache-d23d0d19-4806-4d5e-aa05-13aaf3c4a770" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.097029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Acquired lock "refresh_cache-d23d0d19-4806-4d5e-aa05-13aaf3c4a770" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.097029] env[61906]: DEBUG nova.network.neutron [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 597.299119] env[61906]: INFO nova.compute.manager [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] [instance: 6b3ecea4-fe20-4b6c-a987-59c20731b9ac] Took 1.02 seconds to deallocate network for instance. [ 597.398263] env[61906]: DEBUG nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.624294] env[61906]: DEBUG nova.network.neutron [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.730097] env[61906]: DEBUG nova.network.neutron [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.919356] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.233034] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Releasing lock "refresh_cache-d23d0d19-4806-4d5e-aa05-13aaf3c4a770" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.233665] env[61906]: DEBUG nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 598.233665] env[61906]: DEBUG nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.233809] env[61906]: DEBUG nova.network.neutron [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.267703] env[61906]: DEBUG nova.network.neutron [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.335391] env[61906]: INFO nova.scheduler.client.report [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Deleted allocations for instance 6b3ecea4-fe20-4b6c-a987-59c20731b9ac [ 598.529573] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03655212-81e6-416c-8f45-d3b39edd6013 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.537174] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae87552-f943-4fb6-a6d2-f408a2ac9479 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.567402] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf07b60c-53b9-4f60-b54a-88f8f3d5fc3e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.574957] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03466b8-51d9-453b-b12d-cf886b7687ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.589802] env[61906]: DEBUG nova.compute.provider_tree [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.771513] env[61906]: DEBUG nova.network.neutron [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.845784] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8966e2e-7bab-4cc7-807f-b3a115a37ad8 tempest-ServersTestFqdnHostnames-520220329 tempest-ServersTestFqdnHostnames-520220329-project-member] Lock "6b3ecea4-fe20-4b6c-a987-59c20731b9ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.410s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.097381] env[61906]: DEBUG nova.scheduler.client.report [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.273985] env[61906]: INFO nova.compute.manager [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] [instance: d23d0d19-4806-4d5e-aa05-13aaf3c4a770] Took 1.04 seconds to deallocate network for instance. [ 599.349023] env[61906]: DEBUG nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 599.602820] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.603357] env[61906]: DEBUG nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 599.605948] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.581s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.872176] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.110466] env[61906]: DEBUG nova.compute.utils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.111982] env[61906]: DEBUG nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.112299] env[61906]: DEBUG nova.network.neutron [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 600.202927] env[61906]: DEBUG nova.policy [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '649eabb1fe0e4a4884624e27eade0ee7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cff38ff9380e44e4adc18c532e4ff258', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 600.321243] env[61906]: INFO nova.scheduler.client.report [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Deleted allocations for instance d23d0d19-4806-4d5e-aa05-13aaf3c4a770 [ 600.516305] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd83ba5-7aa1-4a2a-b716-35b24fa7ac13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.523945] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83adf861-ec03-4cdd-8cf9-18f8fc95623e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.573412] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fe49de7-9f60-418c-ae6e-fa962b460ec9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.581777] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384cf0e6-6223-45e5-a4d5-348b37ddddf0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.596217] env[61906]: DEBUG nova.compute.provider_tree [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 600.618984] env[61906]: DEBUG nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 600.742485] env[61906]: DEBUG nova.network.neutron [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Successfully created port: 1879d0be-ab99-41ea-a229-33aba99ee775 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 600.831312] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0fe0db35-8594-4ea5-b20c-d71dbb7eb846 tempest-ServerExternalEventsTest-1914340765 tempest-ServerExternalEventsTest-1914340765-project-member] Lock "d23d0d19-4806-4d5e-aa05-13aaf3c4a770" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.821s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.103016] env[61906]: DEBUG nova.scheduler.client.report [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.335606] env[61906]: DEBUG nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 601.608592] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.003s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.609276] env[61906]: ERROR nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a477cd11-7da0-45e8-8670-8c0022663392, please check neutron logs for more information. [ 601.609276] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Traceback (most recent call last): [ 601.609276] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.609276] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] self.driver.spawn(context, instance, image_meta, [ 601.609276] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 601.609276] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.609276] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.609276] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] vm_ref = self.build_virtual_machine(instance, [ 601.609276] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.609276] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.609276] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.609745] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] for vif in network_info: [ 601.609745] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.609745] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] return self._sync_wrapper(fn, *args, **kwargs) [ 601.609745] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.609745] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] self.wait() [ 601.609745] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.609745] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] self[:] = self._gt.wait() [ 601.609745] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.609745] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] return self._exit_event.wait() [ 601.609745] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.609745] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] result = hub.switch() [ 601.609745] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.609745] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] return self.greenlet.switch() [ 601.610154] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.610154] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] result = function(*args, **kwargs) [ 601.610154] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 601.610154] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] return func(*args, **kwargs) [ 601.610154] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.610154] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] raise e [ 601.610154] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.610154] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] nwinfo = self.network_api.allocate_for_instance( [ 601.610154] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.610154] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] created_port_ids = self._update_ports_for_instance( [ 601.610154] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.610154] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] with excutils.save_and_reraise_exception(): [ 601.610154] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.610440] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] self.force_reraise() [ 601.610440] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.610440] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] raise self.value [ 601.610440] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.610440] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] updated_port = self._update_port( [ 601.610440] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.610440] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] _ensure_no_port_binding_failure(port) [ 601.610440] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.610440] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] raise exception.PortBindingFailed(port_id=port['id']) [ 601.610440] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] nova.exception.PortBindingFailed: Binding failed for port a477cd11-7da0-45e8-8670-8c0022663392, please check neutron logs for more information. [ 601.610440] env[61906]: ERROR nova.compute.manager [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] [ 601.610696] env[61906]: DEBUG nova.compute.utils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Binding failed for port a477cd11-7da0-45e8-8670-8c0022663392, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 601.611859] env[61906]: DEBUG nova.compute.manager [req-e45b77f8-1d38-461c-b135-f46b16dae6db req-ce85f32a-fda6-4088-b9a7-da5d5cea8dcd service nova] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Received event network-changed-1879d0be-ab99-41ea-a229-33aba99ee775 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 601.612042] env[61906]: DEBUG nova.compute.manager [req-e45b77f8-1d38-461c-b135-f46b16dae6db req-ce85f32a-fda6-4088-b9a7-da5d5cea8dcd service nova] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Refreshing instance network info cache due to event network-changed-1879d0be-ab99-41ea-a229-33aba99ee775. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 601.612261] env[61906]: DEBUG oslo_concurrency.lockutils [req-e45b77f8-1d38-461c-b135-f46b16dae6db req-ce85f32a-fda6-4088-b9a7-da5d5cea8dcd service nova] Acquiring lock "refresh_cache-40f3efa6-699f-4779-b999-c86435e04c0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.612406] env[61906]: DEBUG oslo_concurrency.lockutils [req-e45b77f8-1d38-461c-b135-f46b16dae6db req-ce85f32a-fda6-4088-b9a7-da5d5cea8dcd service nova] Acquired lock "refresh_cache-40f3efa6-699f-4779-b999-c86435e04c0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.612565] env[61906]: DEBUG nova.network.neutron [req-e45b77f8-1d38-461c-b135-f46b16dae6db req-ce85f32a-fda6-4088-b9a7-da5d5cea8dcd service nova] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Refreshing network info cache for port 1879d0be-ab99-41ea-a229-33aba99ee775 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 601.613824] env[61906]: DEBUG nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Build of instance 528a2095-e82e-44eb-a4a1-dacde88966bb was re-scheduled: Binding failed for port a477cd11-7da0-45e8-8670-8c0022663392, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 601.614252] env[61906]: DEBUG nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 601.614479] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Acquiring lock "refresh_cache-528a2095-e82e-44eb-a4a1-dacde88966bb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.614634] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Acquired lock "refresh_cache-528a2095-e82e-44eb-a4a1-dacde88966bb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.614789] env[61906]: DEBUG nova.network.neutron [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.615651] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.206s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.617571] env[61906]: INFO nova.compute.claims [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 601.630897] env[61906]: DEBUG nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 601.669648] env[61906]: DEBUG nova.virt.hardware [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 601.669993] env[61906]: DEBUG nova.virt.hardware [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 601.670117] env[61906]: DEBUG nova.virt.hardware [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 601.670333] env[61906]: DEBUG nova.virt.hardware [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 601.670505] env[61906]: DEBUG nova.virt.hardware [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 601.670682] env[61906]: DEBUG nova.virt.hardware [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 601.670961] env[61906]: DEBUG nova.virt.hardware [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 601.671252] env[61906]: DEBUG nova.virt.hardware [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 601.671551] env[61906]: DEBUG nova.virt.hardware [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 601.671767] env[61906]: DEBUG nova.virt.hardware [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 601.671995] env[61906]: DEBUG nova.virt.hardware [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.673339] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08704a02-6893-405f-963a-d29f21a3f2e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.684623] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d6c333-a1bd-4c2e-b507-499b5366f97c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.847876] env[61906]: ERROR nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1879d0be-ab99-41ea-a229-33aba99ee775, please check neutron logs for more information. [ 601.847876] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.847876] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.847876] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.847876] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.847876] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.847876] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.847876] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.847876] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.847876] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 601.847876] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.847876] env[61906]: ERROR nova.compute.manager raise self.value [ 601.847876] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.847876] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.847876] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.847876] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.848349] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.848349] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.848349] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1879d0be-ab99-41ea-a229-33aba99ee775, please check neutron logs for more information. [ 601.848349] env[61906]: ERROR nova.compute.manager [ 601.850209] env[61906]: Traceback (most recent call last): [ 601.850209] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.850209] env[61906]: listener.cb(fileno) [ 601.850209] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.850209] env[61906]: result = function(*args, **kwargs) [ 601.850209] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 601.850209] env[61906]: return func(*args, **kwargs) [ 601.850209] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.850209] env[61906]: raise e [ 601.850209] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.850209] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 601.850209] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.850209] env[61906]: created_port_ids = self._update_ports_for_instance( [ 601.850209] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.850209] env[61906]: with excutils.save_and_reraise_exception(): [ 601.850209] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.850209] env[61906]: self.force_reraise() [ 601.850209] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.850209] env[61906]: raise self.value [ 601.850209] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.850209] env[61906]: updated_port = self._update_port( [ 601.850209] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.850209] env[61906]: _ensure_no_port_binding_failure(port) [ 601.850209] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.850209] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.850209] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 1879d0be-ab99-41ea-a229-33aba99ee775, please check neutron logs for more information. [ 601.850209] env[61906]: Removing descriptor: 15 [ 601.850965] env[61906]: ERROR nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1879d0be-ab99-41ea-a229-33aba99ee775, please check neutron logs for more information. [ 601.850965] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Traceback (most recent call last): [ 601.850965] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 601.850965] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] yield resources [ 601.850965] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.850965] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] self.driver.spawn(context, instance, image_meta, [ 601.850965] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 601.850965] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.850965] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.850965] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] vm_ref = self.build_virtual_machine(instance, [ 601.850965] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.851324] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.851324] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.851324] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] for vif in network_info: [ 601.851324] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.851324] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] return self._sync_wrapper(fn, *args, **kwargs) [ 601.851324] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.851324] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] self.wait() [ 601.851324] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.851324] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] self[:] = self._gt.wait() [ 601.851324] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.851324] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] return self._exit_event.wait() [ 601.851324] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.851324] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] result = hub.switch() [ 601.851669] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.851669] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] return self.greenlet.switch() [ 601.851669] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.851669] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] result = function(*args, **kwargs) [ 601.851669] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 601.851669] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] return func(*args, **kwargs) [ 601.851669] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.851669] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] raise e [ 601.851669] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.851669] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] nwinfo = self.network_api.allocate_for_instance( [ 601.851669] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.851669] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] created_port_ids = self._update_ports_for_instance( [ 601.851669] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.852027] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] with excutils.save_and_reraise_exception(): [ 601.852027] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.852027] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] self.force_reraise() [ 601.852027] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.852027] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] raise self.value [ 601.852027] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.852027] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] updated_port = self._update_port( [ 601.852027] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.852027] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] _ensure_no_port_binding_failure(port) [ 601.852027] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.852027] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] raise exception.PortBindingFailed(port_id=port['id']) [ 601.852027] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] nova.exception.PortBindingFailed: Binding failed for port 1879d0be-ab99-41ea-a229-33aba99ee775, please check neutron logs for more information. [ 601.852027] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] [ 601.852419] env[61906]: INFO nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Terminating instance [ 601.855810] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "refresh_cache-40f3efa6-699f-4779-b999-c86435e04c0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.866968] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.154556] env[61906]: DEBUG nova.network.neutron [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.157086] env[61906]: DEBUG nova.network.neutron [req-e45b77f8-1d38-461c-b135-f46b16dae6db req-ce85f32a-fda6-4088-b9a7-da5d5cea8dcd service nova] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.286597] env[61906]: DEBUG nova.network.neutron [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.316137] env[61906]: DEBUG nova.network.neutron [req-e45b77f8-1d38-461c-b135-f46b16dae6db req-ce85f32a-fda6-4088-b9a7-da5d5cea8dcd service nova] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.789509] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Releasing lock "refresh_cache-528a2095-e82e-44eb-a4a1-dacde88966bb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.789853] env[61906]: DEBUG nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 602.790704] env[61906]: DEBUG nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 602.790704] env[61906]: DEBUG nova.network.neutron [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 602.812854] env[61906]: DEBUG nova.network.neutron [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.820257] env[61906]: DEBUG oslo_concurrency.lockutils [req-e45b77f8-1d38-461c-b135-f46b16dae6db req-ce85f32a-fda6-4088-b9a7-da5d5cea8dcd service nova] Releasing lock "refresh_cache-40f3efa6-699f-4779-b999-c86435e04c0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.820601] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquired lock "refresh_cache-40f3efa6-699f-4779-b999-c86435e04c0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.820782] env[61906]: DEBUG nova.network.neutron [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 603.039481] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa789ad-55c4-4fbb-82a6-f9a39e3bed26 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.047993] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e73eba9-7c5d-42b5-baee-81ceaeea5c8a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.078150] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95785580-ca1a-4843-90fe-ba26bf31c392 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.087648] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e283f8-04de-4491-80be-15686ea75fb2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.100402] env[61906]: DEBUG nova.compute.provider_tree [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.318250] env[61906]: DEBUG nova.network.neutron [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.356596] env[61906]: DEBUG nova.network.neutron [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.487275] env[61906]: DEBUG nova.network.neutron [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.609107] env[61906]: DEBUG nova.scheduler.client.report [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.641160] env[61906]: DEBUG nova.compute.manager [req-a916a0b6-0c07-4ee6-b290-4d9eca15f23a req-c84db214-4611-47d6-8fd6-29d66824bc65 service nova] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Received event network-vif-deleted-1879d0be-ab99-41ea-a229-33aba99ee775 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 603.821512] env[61906]: INFO nova.compute.manager [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] [instance: 528a2095-e82e-44eb-a4a1-dacde88966bb] Took 1.03 seconds to deallocate network for instance. [ 603.992920] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Releasing lock "refresh_cache-40f3efa6-699f-4779-b999-c86435e04c0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.993500] env[61906]: DEBUG nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 603.993637] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 603.993969] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b0bbe52-3af5-4554-a556-eaea99d88b18 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.008623] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956a71c6-4736-48ae-986a-914d52635bc7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.032124] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 40f3efa6-699f-4779-b999-c86435e04c0a could not be found. [ 604.032433] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 604.032685] env[61906]: INFO nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 604.033161] env[61906]: DEBUG oslo.service.loopingcall [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.033584] env[61906]: DEBUG nova.compute.manager [-] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.033768] env[61906]: DEBUG nova.network.neutron [-] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 604.067546] env[61906]: DEBUG nova.network.neutron [-] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.117409] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.117409] env[61906]: DEBUG nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 604.120083] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.938s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.574440] env[61906]: DEBUG nova.network.neutron [-] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.626310] env[61906]: DEBUG nova.compute.utils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 604.631760] env[61906]: DEBUG nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 604.636021] env[61906]: DEBUG nova.network.neutron [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 604.705770] env[61906]: DEBUG nova.policy [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '649eabb1fe0e4a4884624e27eade0ee7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cff38ff9380e44e4adc18c532e4ff258', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 604.861930] env[61906]: INFO nova.scheduler.client.report [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Deleted allocations for instance 528a2095-e82e-44eb-a4a1-dacde88966bb [ 604.882194] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Acquiring lock "b1ec5d55-96a6-42eb-ba54-c4a3904e22cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.888437] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Lock "b1ec5d55-96a6-42eb-ba54-c4a3904e22cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.005s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.082243] env[61906]: INFO nova.compute.manager [-] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Took 1.05 seconds to deallocate network for instance. [ 605.084399] env[61906]: DEBUG nova.compute.claims [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 605.084621] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.123719] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34a7128-1062-4f86-bbb0-9e16956e548d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.131799] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672a75ba-7f57-4b14-8195-1bfaaf325d1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.164334] env[61906]: DEBUG nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 605.169107] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af47907-fef4-4930-b823-036ec44e41fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.177680] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0d7575-3834-4c2b-9a1b-f58f5ce580df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.194650] env[61906]: DEBUG nova.compute.provider_tree [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.212912] env[61906]: DEBUG nova.network.neutron [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Successfully created port: d85e5455-62ad-4a39-bb2f-c0f60076528c {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 605.374767] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a07f4a3-6347-4469-977b-c1c67047d7d4 tempest-ServerRescueTestJSON-1137876203 tempest-ServerRescueTestJSON-1137876203-project-member] Lock "528a2095-e82e-44eb-a4a1-dacde88966bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.754s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.697781] env[61906]: DEBUG nova.scheduler.client.report [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 605.879650] env[61906]: DEBUG nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 606.175965] env[61906]: DEBUG nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 606.202682] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.082s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.205340] env[61906]: ERROR nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 509913ec-88cc-40e8-a970-b3a19be91b8b, please check neutron logs for more information. [ 606.205340] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Traceback (most recent call last): [ 606.205340] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.205340] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] self.driver.spawn(context, instance, image_meta, [ 606.205340] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 606.205340] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.205340] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.205340] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] vm_ref = self.build_virtual_machine(instance, [ 606.205340] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.205340] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.205340] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.205800] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] for vif in network_info: [ 606.205800] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.205800] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] return self._sync_wrapper(fn, *args, **kwargs) [ 606.205800] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.205800] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] self.wait() [ 606.205800] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.205800] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] self[:] = self._gt.wait() [ 606.205800] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.205800] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] return self._exit_event.wait() [ 606.205800] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.205800] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] result = hub.switch() [ 606.205800] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.205800] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] return self.greenlet.switch() [ 606.206154] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.206154] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] result = function(*args, **kwargs) [ 606.206154] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.206154] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] return func(*args, **kwargs) [ 606.206154] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.206154] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] raise e [ 606.206154] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.206154] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] nwinfo = self.network_api.allocate_for_instance( [ 606.206154] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.206154] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] created_port_ids = self._update_ports_for_instance( [ 606.206154] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.206154] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] with excutils.save_and_reraise_exception(): [ 606.206154] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.206726] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] self.force_reraise() [ 606.206726] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.206726] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] raise self.value [ 606.206726] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.206726] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] updated_port = self._update_port( [ 606.206726] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.206726] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] _ensure_no_port_binding_failure(port) [ 606.206726] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.206726] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] raise exception.PortBindingFailed(port_id=port['id']) [ 606.206726] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] nova.exception.PortBindingFailed: Binding failed for port 509913ec-88cc-40e8-a970-b3a19be91b8b, please check neutron logs for more information. [ 606.206726] env[61906]: ERROR nova.compute.manager [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] [ 606.207240] env[61906]: DEBUG nova.compute.utils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Binding failed for port 509913ec-88cc-40e8-a970-b3a19be91b8b, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 606.210799] env[61906]: DEBUG nova.virt.hardware [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 606.211039] env[61906]: DEBUG nova.virt.hardware [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 606.211204] env[61906]: DEBUG nova.virt.hardware [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 606.211605] env[61906]: DEBUG nova.virt.hardware [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 606.211768] env[61906]: DEBUG nova.virt.hardware [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 606.212449] env[61906]: DEBUG nova.virt.hardware [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 606.212737] env[61906]: DEBUG nova.virt.hardware [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 606.213692] env[61906]: DEBUG nova.virt.hardware [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 606.213938] env[61906]: DEBUG nova.virt.hardware [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 606.214207] env[61906]: DEBUG nova.virt.hardware [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 606.215163] env[61906]: DEBUG nova.virt.hardware [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 606.215163] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.890s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.216139] env[61906]: INFO nova.compute.claims [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.219161] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ed4b86-4827-4e1c-b95e-443cd52012a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.227532] env[61906]: DEBUG nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Build of instance e64fd277-10ae-494d-918a-2f3faa4e44d1 was re-scheduled: Binding failed for port 509913ec-88cc-40e8-a970-b3a19be91b8b, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 606.227532] env[61906]: DEBUG nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 606.227532] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquiring lock "refresh_cache-e64fd277-10ae-494d-918a-2f3faa4e44d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.229728] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Acquired lock "refresh_cache-e64fd277-10ae-494d-918a-2f3faa4e44d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.230133] env[61906]: DEBUG nova.network.neutron [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.244760] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fede0f61-a174-4c2e-a8df-23be91802b1f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.409674] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.629191] env[61906]: ERROR nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d85e5455-62ad-4a39-bb2f-c0f60076528c, please check neutron logs for more information. [ 606.629191] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.629191] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.629191] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 606.629191] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.629191] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 606.629191] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.629191] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 606.629191] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.629191] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 606.629191] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.629191] env[61906]: ERROR nova.compute.manager raise self.value [ 606.629191] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.629191] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 606.629191] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.629191] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 606.629580] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.629580] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 606.629580] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d85e5455-62ad-4a39-bb2f-c0f60076528c, please check neutron logs for more information. [ 606.629580] env[61906]: ERROR nova.compute.manager [ 606.629580] env[61906]: Traceback (most recent call last): [ 606.629580] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 606.629580] env[61906]: listener.cb(fileno) [ 606.629580] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.629580] env[61906]: result = function(*args, **kwargs) [ 606.629580] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.629580] env[61906]: return func(*args, **kwargs) [ 606.629580] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.629580] env[61906]: raise e [ 606.629580] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.629580] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 606.629580] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.629580] env[61906]: created_port_ids = self._update_ports_for_instance( [ 606.629580] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.629580] env[61906]: with excutils.save_and_reraise_exception(): [ 606.629580] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.629580] env[61906]: self.force_reraise() [ 606.629580] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.629580] env[61906]: raise self.value [ 606.629580] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.629580] env[61906]: updated_port = self._update_port( [ 606.629580] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.629580] env[61906]: _ensure_no_port_binding_failure(port) [ 606.629580] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.629580] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 606.630220] env[61906]: nova.exception.PortBindingFailed: Binding failed for port d85e5455-62ad-4a39-bb2f-c0f60076528c, please check neutron logs for more information. [ 606.630220] env[61906]: Removing descriptor: 17 [ 606.630220] env[61906]: ERROR nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d85e5455-62ad-4a39-bb2f-c0f60076528c, please check neutron logs for more information. [ 606.630220] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Traceback (most recent call last): [ 606.630220] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 606.630220] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] yield resources [ 606.630220] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.630220] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] self.driver.spawn(context, instance, image_meta, [ 606.630220] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 606.630220] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.630220] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.630220] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] vm_ref = self.build_virtual_machine(instance, [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] for vif in network_info: [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] return self._sync_wrapper(fn, *args, **kwargs) [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] self.wait() [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] self[:] = self._gt.wait() [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] return self._exit_event.wait() [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.630495] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] result = hub.switch() [ 606.630801] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.630801] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] return self.greenlet.switch() [ 606.630801] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.630801] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] result = function(*args, **kwargs) [ 606.630801] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.630801] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] return func(*args, **kwargs) [ 606.630801] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.630801] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] raise e [ 606.630801] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.630801] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] nwinfo = self.network_api.allocate_for_instance( [ 606.630801] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.630801] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] created_port_ids = self._update_ports_for_instance( [ 606.630801] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.631102] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] with excutils.save_and_reraise_exception(): [ 606.631102] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.631102] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] self.force_reraise() [ 606.631102] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.631102] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] raise self.value [ 606.631102] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.631102] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] updated_port = self._update_port( [ 606.631102] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.631102] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] _ensure_no_port_binding_failure(port) [ 606.631102] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.631102] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] raise exception.PortBindingFailed(port_id=port['id']) [ 606.631102] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] nova.exception.PortBindingFailed: Binding failed for port d85e5455-62ad-4a39-bb2f-c0f60076528c, please check neutron logs for more information. [ 606.631102] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] [ 606.631396] env[61906]: INFO nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Terminating instance [ 606.633659] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "refresh_cache-081d06b3-c2b6-42f2-94b9-925493590e55" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.633659] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquired lock "refresh_cache-081d06b3-c2b6-42f2-94b9-925493590e55" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.633659] env[61906]: DEBUG nova.network.neutron [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.755176] env[61906]: DEBUG nova.network.neutron [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.761501] env[61906]: DEBUG nova.compute.manager [req-afd5e0fc-1cd1-437e-ae56-2e2d6861d8e1 req-aefd5e10-d59e-4315-b27d-87aef8420b35 service nova] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Received event network-changed-d85e5455-62ad-4a39-bb2f-c0f60076528c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 606.761501] env[61906]: DEBUG nova.compute.manager [req-afd5e0fc-1cd1-437e-ae56-2e2d6861d8e1 req-aefd5e10-d59e-4315-b27d-87aef8420b35 service nova] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Refreshing instance network info cache due to event network-changed-d85e5455-62ad-4a39-bb2f-c0f60076528c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 606.761501] env[61906]: DEBUG oslo_concurrency.lockutils [req-afd5e0fc-1cd1-437e-ae56-2e2d6861d8e1 req-aefd5e10-d59e-4315-b27d-87aef8420b35 service nova] Acquiring lock "refresh_cache-081d06b3-c2b6-42f2-94b9-925493590e55" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.845673] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Acquiring lock "83c86407-16b6-4025-9568-0c9c73d20b10" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.845997] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Lock "83c86407-16b6-4025-9568-0c9c73d20b10" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.872330] env[61906]: DEBUG nova.network.neutron [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.162068] env[61906]: DEBUG nova.network.neutron [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.286572] env[61906]: DEBUG nova.network.neutron [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.374789] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Releasing lock "refresh_cache-e64fd277-10ae-494d-918a-2f3faa4e44d1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.375079] env[61906]: DEBUG nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 607.375267] env[61906]: DEBUG nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.375430] env[61906]: DEBUG nova.network.neutron [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.407902] env[61906]: DEBUG nova.network.neutron [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.416535] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.418293] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.768493] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19feb5f7-0956-4990-9e1c-66cfb7739055 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.777641] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb35989-4bb2-4e69-a7ca-ae182ae2419c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.810479] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Releasing lock "refresh_cache-081d06b3-c2b6-42f2-94b9-925493590e55" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.811364] env[61906]: DEBUG nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 607.812095] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 607.812687] env[61906]: DEBUG oslo_concurrency.lockutils [req-afd5e0fc-1cd1-437e-ae56-2e2d6861d8e1 req-aefd5e10-d59e-4315-b27d-87aef8420b35 service nova] Acquired lock "refresh_cache-081d06b3-c2b6-42f2-94b9-925493590e55" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.812990] env[61906]: DEBUG nova.network.neutron [req-afd5e0fc-1cd1-437e-ae56-2e2d6861d8e1 req-aefd5e10-d59e-4315-b27d-87aef8420b35 service nova] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Refreshing network info cache for port d85e5455-62ad-4a39-bb2f-c0f60076528c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 607.814383] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b36b90f-7073-490a-a221-c0a57ee76c7c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.820988] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a7ddae-db75-437a-9c90-937e4f3958c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.830066] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef2af5b-828e-4619-8c67-ceb401c8167f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.836624] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8cd105d-e312-4db7-ba21-70bccca93425 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.855940] env[61906]: DEBUG nova.compute.provider_tree [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.861193] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 081d06b3-c2b6-42f2-94b9-925493590e55 could not be found. [ 607.861462] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 607.861462] env[61906]: INFO nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Took 0.05 seconds to destroy the instance on the hypervisor. [ 607.861726] env[61906]: DEBUG oslo.service.loopingcall [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.862117] env[61906]: DEBUG nova.compute.manager [-] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.862210] env[61906]: DEBUG nova.network.neutron [-] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.884087] env[61906]: DEBUG nova.network.neutron [-] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.914797] env[61906]: DEBUG nova.network.neutron [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.929053] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 607.929053] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 607.929053] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 608.344944] env[61906]: DEBUG nova.network.neutron [req-afd5e0fc-1cd1-437e-ae56-2e2d6861d8e1 req-aefd5e10-d59e-4315-b27d-87aef8420b35 service nova] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.363911] env[61906]: DEBUG nova.scheduler.client.report [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.386031] env[61906]: DEBUG nova.network.neutron [-] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.416850] env[61906]: INFO nova.compute.manager [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] [instance: e64fd277-10ae-494d-918a-2f3faa4e44d1] Took 1.04 seconds to deallocate network for instance. [ 608.435313] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 608.435313] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 608.435313] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 608.435313] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 608.435313] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 608.435528] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Didn't find any instances for network info cache update. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 608.438110] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.438110] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.438110] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.438110] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.438110] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.438110] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.438464] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 608.438464] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.454923] env[61906]: DEBUG nova.network.neutron [req-afd5e0fc-1cd1-437e-ae56-2e2d6861d8e1 req-aefd5e10-d59e-4315-b27d-87aef8420b35 service nova] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.868931] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.654s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.869782] env[61906]: DEBUG nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 608.872099] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.025s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.888531] env[61906]: INFO nova.compute.manager [-] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Took 1.03 seconds to deallocate network for instance. [ 608.890828] env[61906]: DEBUG nova.compute.claims [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 608.890828] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.918813] env[61906]: DEBUG nova.compute.manager [req-9d0d756d-434d-41ce-8e06-61b2ca10e1ee req-eceae909-3a9d-412e-a05c-40303be58ef3 service nova] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Received event network-vif-deleted-d85e5455-62ad-4a39-bb2f-c0f60076528c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.939726] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.958130] env[61906]: DEBUG oslo_concurrency.lockutils [req-afd5e0fc-1cd1-437e-ae56-2e2d6861d8e1 req-aefd5e10-d59e-4315-b27d-87aef8420b35 service nova] Releasing lock "refresh_cache-081d06b3-c2b6-42f2-94b9-925493590e55" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.380053] env[61906]: DEBUG nova.compute.utils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 609.382722] env[61906]: DEBUG nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 609.383098] env[61906]: DEBUG nova.network.neutron [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 609.441745] env[61906]: INFO nova.scheduler.client.report [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Deleted allocations for instance e64fd277-10ae-494d-918a-2f3faa4e44d1 [ 609.461520] env[61906]: DEBUG nova.policy [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e40421895a4346638427b8531bbb4dcb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d018a3a09804d6cbaf59c5924fdbe61', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 609.849059] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999cfd59-eb4d-40cf-9a5e-bfbde431efca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.858208] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf31035-be79-4e01-91e4-c263668faf39 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.892615] env[61906]: DEBUG nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 609.897267] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3130adfb-b936-4998-89b2-2005da684135 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.906358] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b9e634-67b6-438f-bfb0-d4c3539a2bda {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.923910] env[61906]: DEBUG nova.compute.provider_tree [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.948824] env[61906]: DEBUG oslo_concurrency.lockutils [None req-82873558-d2cc-43c8-89c1-26d08fb71a06 tempest-DeleteServersAdminTestJSON-1937943042 tempest-DeleteServersAdminTestJSON-1937943042-project-member] Lock "e64fd277-10ae-494d-918a-2f3faa4e44d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.020s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.105221] env[61906]: DEBUG nova.network.neutron [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Successfully created port: 7df4399b-0ff5-4b9b-94b8-ecc1547ee394 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 610.431249] env[61906]: DEBUG nova.scheduler.client.report [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.452858] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 610.911281] env[61906]: DEBUG nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 610.936485] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.064s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.936916] env[61906]: ERROR nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9f9923e9-da89-45a9-9699-906de88dc362, please check neutron logs for more information. [ 610.936916] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Traceback (most recent call last): [ 610.936916] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.936916] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] self.driver.spawn(context, instance, image_meta, [ 610.936916] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.936916] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.936916] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.936916] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] vm_ref = self.build_virtual_machine(instance, [ 610.936916] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.936916] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.936916] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.938192] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] for vif in network_info: [ 610.938192] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.938192] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] return self._sync_wrapper(fn, *args, **kwargs) [ 610.938192] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.938192] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] self.wait() [ 610.938192] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.938192] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] self[:] = self._gt.wait() [ 610.938192] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.938192] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] return self._exit_event.wait() [ 610.938192] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.938192] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] result = hub.switch() [ 610.938192] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.938192] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] return self.greenlet.switch() [ 610.938490] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.938490] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] result = function(*args, **kwargs) [ 610.938490] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.938490] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] return func(*args, **kwargs) [ 610.938490] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.938490] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] raise e [ 610.938490] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.938490] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] nwinfo = self.network_api.allocate_for_instance( [ 610.938490] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.938490] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] created_port_ids = self._update_ports_for_instance( [ 610.938490] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.938490] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] with excutils.save_and_reraise_exception(): [ 610.938490] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.939245] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] self.force_reraise() [ 610.939245] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.939245] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] raise self.value [ 610.939245] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.939245] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] updated_port = self._update_port( [ 610.939245] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.939245] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] _ensure_no_port_binding_failure(port) [ 610.939245] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.939245] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] raise exception.PortBindingFailed(port_id=port['id']) [ 610.939245] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] nova.exception.PortBindingFailed: Binding failed for port 9f9923e9-da89-45a9-9699-906de88dc362, please check neutron logs for more information. [ 610.939245] env[61906]: ERROR nova.compute.manager [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] [ 610.939520] env[61906]: DEBUG nova.compute.utils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Binding failed for port 9f9923e9-da89-45a9-9699-906de88dc362, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 610.939520] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.119s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.941933] env[61906]: DEBUG nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Build of instance 4b0abc14-0ce2-4212-8677-82f6279ee107 was re-scheduled: Binding failed for port 9f9923e9-da89-45a9-9699-906de88dc362, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 610.947556] env[61906]: DEBUG nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 610.947556] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Acquiring lock "refresh_cache-4b0abc14-0ce2-4212-8677-82f6279ee107" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.947556] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Acquired lock "refresh_cache-4b0abc14-0ce2-4212-8677-82f6279ee107" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.947556] env[61906]: DEBUG nova.network.neutron [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.964885] env[61906]: DEBUG nova.virt.hardware [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.965146] env[61906]: DEBUG nova.virt.hardware [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.965470] env[61906]: DEBUG nova.virt.hardware [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.965661] env[61906]: DEBUG nova.virt.hardware [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.966177] env[61906]: DEBUG nova.virt.hardware [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.966177] env[61906]: DEBUG nova.virt.hardware [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.966177] env[61906]: DEBUG nova.virt.hardware [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.966360] env[61906]: DEBUG nova.virt.hardware [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.966482] env[61906]: DEBUG nova.virt.hardware [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.966645] env[61906]: DEBUG nova.virt.hardware [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.966888] env[61906]: DEBUG nova.virt.hardware [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.969803] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5859c94-4db4-47ff-96c7-3db3d5a8ff70 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.988367] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc3d43a-4687-4327-bb3b-c6bac4f81ab4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.995222] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.332352] env[61906]: DEBUG nova.compute.manager [req-eb36ad0f-e5bc-4cc6-8562-31cc533b77ba req-5957a0d8-3251-4fb5-802b-259a565c9517 service nova] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Received event network-changed-7df4399b-0ff5-4b9b-94b8-ecc1547ee394 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 611.332913] env[61906]: DEBUG nova.compute.manager [req-eb36ad0f-e5bc-4cc6-8562-31cc533b77ba req-5957a0d8-3251-4fb5-802b-259a565c9517 service nova] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Refreshing instance network info cache due to event network-changed-7df4399b-0ff5-4b9b-94b8-ecc1547ee394. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 611.333236] env[61906]: DEBUG oslo_concurrency.lockutils [req-eb36ad0f-e5bc-4cc6-8562-31cc533b77ba req-5957a0d8-3251-4fb5-802b-259a565c9517 service nova] Acquiring lock "refresh_cache-ebea667f-6495-41c0-88b3-a9c61390558d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.333816] env[61906]: DEBUG oslo_concurrency.lockutils [req-eb36ad0f-e5bc-4cc6-8562-31cc533b77ba req-5957a0d8-3251-4fb5-802b-259a565c9517 service nova] Acquired lock "refresh_cache-ebea667f-6495-41c0-88b3-a9c61390558d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.334079] env[61906]: DEBUG nova.network.neutron [req-eb36ad0f-e5bc-4cc6-8562-31cc533b77ba req-5957a0d8-3251-4fb5-802b-259a565c9517 service nova] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Refreshing network info cache for port 7df4399b-0ff5-4b9b-94b8-ecc1547ee394 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 611.470957] env[61906]: DEBUG nova.network.neutron [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.563401] env[61906]: DEBUG nova.network.neutron [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.568149] env[61906]: ERROR nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7df4399b-0ff5-4b9b-94b8-ecc1547ee394, please check neutron logs for more information. [ 611.568149] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.568149] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.568149] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.568149] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.568149] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.568149] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.568149] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.568149] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.568149] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 611.568149] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.568149] env[61906]: ERROR nova.compute.manager raise self.value [ 611.568149] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.568149] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.568149] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.568149] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.568801] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.568801] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.568801] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7df4399b-0ff5-4b9b-94b8-ecc1547ee394, please check neutron logs for more information. [ 611.568801] env[61906]: ERROR nova.compute.manager [ 611.568801] env[61906]: Traceback (most recent call last): [ 611.568801] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.568801] env[61906]: listener.cb(fileno) [ 611.568801] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.568801] env[61906]: result = function(*args, **kwargs) [ 611.568801] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.568801] env[61906]: return func(*args, **kwargs) [ 611.568801] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.568801] env[61906]: raise e [ 611.568801] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.568801] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 611.568801] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.568801] env[61906]: created_port_ids = self._update_ports_for_instance( [ 611.568801] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.568801] env[61906]: with excutils.save_and_reraise_exception(): [ 611.568801] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.568801] env[61906]: self.force_reraise() [ 611.568801] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.568801] env[61906]: raise self.value [ 611.568801] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.568801] env[61906]: updated_port = self._update_port( [ 611.568801] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.568801] env[61906]: _ensure_no_port_binding_failure(port) [ 611.568801] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.568801] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.569482] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 7df4399b-0ff5-4b9b-94b8-ecc1547ee394, please check neutron logs for more information. [ 611.569482] env[61906]: Removing descriptor: 17 [ 611.569482] env[61906]: ERROR nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7df4399b-0ff5-4b9b-94b8-ecc1547ee394, please check neutron logs for more information. [ 611.569482] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Traceback (most recent call last): [ 611.569482] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 611.569482] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] yield resources [ 611.569482] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.569482] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] self.driver.spawn(context, instance, image_meta, [ 611.569482] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 611.569482] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.569482] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.569482] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] vm_ref = self.build_virtual_machine(instance, [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] for vif in network_info: [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] return self._sync_wrapper(fn, *args, **kwargs) [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] self.wait() [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] self[:] = self._gt.wait() [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] return self._exit_event.wait() [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.569848] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] result = hub.switch() [ 611.570190] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.570190] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] return self.greenlet.switch() [ 611.570190] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.570190] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] result = function(*args, **kwargs) [ 611.570190] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.570190] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] return func(*args, **kwargs) [ 611.570190] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.570190] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] raise e [ 611.570190] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.570190] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] nwinfo = self.network_api.allocate_for_instance( [ 611.570190] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.570190] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] created_port_ids = self._update_ports_for_instance( [ 611.570190] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.570503] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] with excutils.save_and_reraise_exception(): [ 611.570503] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.570503] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] self.force_reraise() [ 611.570503] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.570503] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] raise self.value [ 611.570503] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.570503] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] updated_port = self._update_port( [ 611.570503] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.570503] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] _ensure_no_port_binding_failure(port) [ 611.570503] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.570503] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] raise exception.PortBindingFailed(port_id=port['id']) [ 611.570503] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] nova.exception.PortBindingFailed: Binding failed for port 7df4399b-0ff5-4b9b-94b8-ecc1547ee394, please check neutron logs for more information. [ 611.570503] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] [ 611.570826] env[61906]: INFO nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Terminating instance [ 611.575417] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Acquiring lock "refresh_cache-ebea667f-6495-41c0-88b3-a9c61390558d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.856291] env[61906]: DEBUG nova.network.neutron [req-eb36ad0f-e5bc-4cc6-8562-31cc533b77ba req-5957a0d8-3251-4fb5-802b-259a565c9517 service nova] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.872757] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf64deba-bb1e-4235-8578-d51dadd3a227 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.886822] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48ba98c4-fc80-4b47-b5a9-3ee73a55c98a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.922838] env[61906]: DEBUG nova.network.neutron [req-eb36ad0f-e5bc-4cc6-8562-31cc533b77ba req-5957a0d8-3251-4fb5-802b-259a565c9517 service nova] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.922838] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f1b41ad-e0e3-4659-acac-f3c0e3f4b46d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.930103] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d146fd9c-8653-493c-a69a-dba6561d2468 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.944937] env[61906]: DEBUG nova.compute.provider_tree [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.066673] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Releasing lock "refresh_cache-4b0abc14-0ce2-4212-8677-82f6279ee107" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.066937] env[61906]: DEBUG nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 612.067145] env[61906]: DEBUG nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.067315] env[61906]: DEBUG nova.network.neutron [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.095066] env[61906]: DEBUG nova.network.neutron [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.426741] env[61906]: DEBUG oslo_concurrency.lockutils [req-eb36ad0f-e5bc-4cc6-8562-31cc533b77ba req-5957a0d8-3251-4fb5-802b-259a565c9517 service nova] Releasing lock "refresh_cache-ebea667f-6495-41c0-88b3-a9c61390558d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.427287] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Acquired lock "refresh_cache-ebea667f-6495-41c0-88b3-a9c61390558d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.427637] env[61906]: DEBUG nova.network.neutron [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.448236] env[61906]: DEBUG nova.scheduler.client.report [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.598344] env[61906]: DEBUG nova.network.neutron [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.952580] env[61906]: DEBUG nova.network.neutron [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.959017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.018s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.959017] env[61906]: ERROR nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 49eae8f7-1b56-4377-9e75-390f8c019ca6, please check neutron logs for more information. [ 612.959017] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Traceback (most recent call last): [ 612.959017] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 612.959017] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] self.driver.spawn(context, instance, image_meta, [ 612.959017] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 612.959017] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.959017] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.959017] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] vm_ref = self.build_virtual_machine(instance, [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] for vif in network_info: [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] return self._sync_wrapper(fn, *args, **kwargs) [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] self.wait() [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] self[:] = self._gt.wait() [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] return self._exit_event.wait() [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 612.959347] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] result = hub.switch() [ 612.959743] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 612.959743] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] return self.greenlet.switch() [ 612.959743] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.959743] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] result = function(*args, **kwargs) [ 612.959743] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 612.959743] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] return func(*args, **kwargs) [ 612.959743] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.959743] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] raise e [ 612.959743] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.959743] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] nwinfo = self.network_api.allocate_for_instance( [ 612.959743] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.959743] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] created_port_ids = self._update_ports_for_instance( [ 612.959743] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.960112] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] with excutils.save_and_reraise_exception(): [ 612.960112] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.960112] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] self.force_reraise() [ 612.960112] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.960112] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] raise self.value [ 612.960112] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.960112] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] updated_port = self._update_port( [ 612.960112] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.960112] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] _ensure_no_port_binding_failure(port) [ 612.960112] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.960112] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] raise exception.PortBindingFailed(port_id=port['id']) [ 612.960112] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] nova.exception.PortBindingFailed: Binding failed for port 49eae8f7-1b56-4377-9e75-390f8c019ca6, please check neutron logs for more information. [ 612.960112] env[61906]: ERROR nova.compute.manager [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] [ 612.960531] env[61906]: DEBUG nova.compute.utils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Binding failed for port 49eae8f7-1b56-4377-9e75-390f8c019ca6, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 612.961081] env[61906]: DEBUG nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Build of instance 05cefc82-3585-4d0b-8620-91412b3ef69e was re-scheduled: Binding failed for port 49eae8f7-1b56-4377-9e75-390f8c019ca6, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 612.961721] env[61906]: DEBUG nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 612.962057] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Acquiring lock "refresh_cache-05cefc82-3585-4d0b-8620-91412b3ef69e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.962310] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Acquired lock "refresh_cache-05cefc82-3585-4d0b-8620-91412b3ef69e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.962598] env[61906]: DEBUG nova.network.neutron [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.963754] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.045s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.965456] env[61906]: INFO nova.compute.claims [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.009409] env[61906]: DEBUG nova.network.neutron [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.101678] env[61906]: INFO nova.compute.manager [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] [instance: 4b0abc14-0ce2-4212-8677-82f6279ee107] Took 1.03 seconds to deallocate network for instance. [ 613.361571] env[61906]: DEBUG nova.compute.manager [req-044fa71f-98c6-486e-8995-2761a2b83315 req-810e75a4-d12f-4d0f-a164-1de690e6144e service nova] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Received event network-vif-deleted-7df4399b-0ff5-4b9b-94b8-ecc1547ee394 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 613.491977] env[61906]: DEBUG nova.network.neutron [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.512047] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Releasing lock "refresh_cache-ebea667f-6495-41c0-88b3-a9c61390558d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.512491] env[61906]: DEBUG nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 613.512691] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 613.512991] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aed36a99-2bc8-4bcb-8a08-489b7b90d6bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.523128] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c99fd40-acd5-4c15-9a24-435de16f3ba4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.547879] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ebea667f-6495-41c0-88b3-a9c61390558d could not be found. [ 613.548166] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 613.548396] env[61906]: INFO nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 613.548660] env[61906]: DEBUG oslo.service.loopingcall [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.548865] env[61906]: DEBUG nova.compute.manager [-] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.548959] env[61906]: DEBUG nova.network.neutron [-] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.578280] env[61906]: DEBUG nova.network.neutron [-] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.597742] env[61906]: DEBUG nova.network.neutron [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.084027] env[61906]: DEBUG nova.network.neutron [-] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.101830] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Releasing lock "refresh_cache-05cefc82-3585-4d0b-8620-91412b3ef69e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.101830] env[61906]: DEBUG nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 614.101830] env[61906]: DEBUG nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.102268] env[61906]: DEBUG nova.network.neutron [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.125892] env[61906]: DEBUG nova.network.neutron [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.143490] env[61906]: INFO nova.scheduler.client.report [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Deleted allocations for instance 4b0abc14-0ce2-4212-8677-82f6279ee107 [ 614.438993] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1494f82-454a-4625-b927-2dcd4330a8cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.446621] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43aa1206-658a-4d3d-a3d4-e2683603a22e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.478471] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a97084d-8a0a-4b41-98af-5760c26afc13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.484627] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100109ba-4230-427b-bbe6-cd046209dfe8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.497854] env[61906]: DEBUG nova.compute.provider_tree [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.586778] env[61906]: INFO nova.compute.manager [-] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Took 1.04 seconds to deallocate network for instance. [ 614.590252] env[61906]: DEBUG nova.compute.claims [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 614.590478] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.631185] env[61906]: DEBUG nova.network.neutron [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.652919] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3653ef56-9b83-43ec-90f8-ed09196a8ae6 tempest-ImagesOneServerNegativeTestJSON-1875036442 tempest-ImagesOneServerNegativeTestJSON-1875036442-project-member] Lock "4b0abc14-0ce2-4212-8677-82f6279ee107" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.364s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.001660] env[61906]: DEBUG nova.scheduler.client.report [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.134801] env[61906]: INFO nova.compute.manager [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] [instance: 05cefc82-3585-4d0b-8620-91412b3ef69e] Took 1.03 seconds to deallocate network for instance. [ 615.155376] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 615.509668] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.510220] env[61906]: DEBUG nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 615.513705] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.641s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.516056] env[61906]: INFO nova.compute.claims [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 615.688651] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.024950] env[61906]: DEBUG nova.compute.utils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 616.030195] env[61906]: DEBUG nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 616.030195] env[61906]: DEBUG nova.network.neutron [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 616.101216] env[61906]: DEBUG nova.policy [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '649eabb1fe0e4a4884624e27eade0ee7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cff38ff9380e44e4adc18c532e4ff258', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 616.180058] env[61906]: INFO nova.scheduler.client.report [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Deleted allocations for instance 05cefc82-3585-4d0b-8620-91412b3ef69e [ 616.531820] env[61906]: DEBUG nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 616.615267] env[61906]: DEBUG nova.network.neutron [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Successfully created port: 9aab81df-f193-4ebc-a8ca-e4b8892b9597 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 616.693117] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c28ec02b-f5b6-4ee7-acc9-81bd2b3c7a69 tempest-ServerActionsTestOtherB-468575615 tempest-ServerActionsTestOtherB-468575615-project-member] Lock "05cefc82-3585-4d0b-8620-91412b3ef69e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.620s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.959434] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4678893d-db12-4d1e-aa6c-959ad20f5436 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.967919] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b435ca47-b388-4b8d-a2b5-16ccf37ab3bb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.002963] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d624bd-830d-4435-b5dc-85768580618e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.014023] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efaf6a0-89aa-488e-9f27-238c40c98fb8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.030145] env[61906]: DEBUG nova.compute.provider_tree [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.197813] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 617.533840] env[61906]: DEBUG nova.scheduler.client.report [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.548437] env[61906]: DEBUG nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 617.588086] env[61906]: DEBUG nova.virt.hardware [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 617.588375] env[61906]: DEBUG nova.virt.hardware [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 617.588542] env[61906]: DEBUG nova.virt.hardware [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 617.589230] env[61906]: DEBUG nova.virt.hardware [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 617.589230] env[61906]: DEBUG nova.virt.hardware [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 617.589355] env[61906]: DEBUG nova.virt.hardware [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 617.592054] env[61906]: DEBUG nova.virt.hardware [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 617.592054] env[61906]: DEBUG nova.virt.hardware [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 617.592054] env[61906]: DEBUG nova.virt.hardware [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 617.592054] env[61906]: DEBUG nova.virt.hardware [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 617.592054] env[61906]: DEBUG nova.virt.hardware [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 617.592272] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25ee3a6-2b41-4d73-8a1c-f25b94768002 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.601057] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b91d6c6-ccf2-47bf-92ae-48509bd69141 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.722809] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.881301] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Acquiring lock "2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.883530] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Lock "2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.040684] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.527s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.041035] env[61906]: DEBUG nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 618.043776] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.177s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.045166] env[61906]: INFO nova.compute.claims [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 618.095366] env[61906]: DEBUG nova.compute.manager [req-7464899d-3b68-429e-afbd-803c091aa8ab req-5664214f-d726-40d7-a840-b31b74e634ed service nova] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Received event network-changed-9aab81df-f193-4ebc-a8ca-e4b8892b9597 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.095565] env[61906]: DEBUG nova.compute.manager [req-7464899d-3b68-429e-afbd-803c091aa8ab req-5664214f-d726-40d7-a840-b31b74e634ed service nova] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Refreshing instance network info cache due to event network-changed-9aab81df-f193-4ebc-a8ca-e4b8892b9597. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 618.095781] env[61906]: DEBUG oslo_concurrency.lockutils [req-7464899d-3b68-429e-afbd-803c091aa8ab req-5664214f-d726-40d7-a840-b31b74e634ed service nova] Acquiring lock "refresh_cache-b56dd330-d31c-4e42-baf9-d9897c55e633" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.096244] env[61906]: DEBUG oslo_concurrency.lockutils [req-7464899d-3b68-429e-afbd-803c091aa8ab req-5664214f-d726-40d7-a840-b31b74e634ed service nova] Acquired lock "refresh_cache-b56dd330-d31c-4e42-baf9-d9897c55e633" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.096244] env[61906]: DEBUG nova.network.neutron [req-7464899d-3b68-429e-afbd-803c091aa8ab req-5664214f-d726-40d7-a840-b31b74e634ed service nova] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Refreshing network info cache for port 9aab81df-f193-4ebc-a8ca-e4b8892b9597 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 618.504174] env[61906]: ERROR nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9aab81df-f193-4ebc-a8ca-e4b8892b9597, please check neutron logs for more information. [ 618.504174] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 618.504174] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.504174] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 618.504174] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.504174] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 618.504174] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.504174] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 618.504174] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.504174] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 618.504174] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.504174] env[61906]: ERROR nova.compute.manager raise self.value [ 618.504174] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.504174] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 618.504174] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.504174] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 618.504897] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.504897] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 618.504897] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9aab81df-f193-4ebc-a8ca-e4b8892b9597, please check neutron logs for more information. [ 618.504897] env[61906]: ERROR nova.compute.manager [ 618.504897] env[61906]: Traceback (most recent call last): [ 618.504897] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 618.504897] env[61906]: listener.cb(fileno) [ 618.504897] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.504897] env[61906]: result = function(*args, **kwargs) [ 618.504897] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.504897] env[61906]: return func(*args, **kwargs) [ 618.504897] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.504897] env[61906]: raise e [ 618.504897] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.504897] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 618.504897] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.504897] env[61906]: created_port_ids = self._update_ports_for_instance( [ 618.504897] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.504897] env[61906]: with excutils.save_and_reraise_exception(): [ 618.504897] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.504897] env[61906]: self.force_reraise() [ 618.504897] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.504897] env[61906]: raise self.value [ 618.504897] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.504897] env[61906]: updated_port = self._update_port( [ 618.504897] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.504897] env[61906]: _ensure_no_port_binding_failure(port) [ 618.504897] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.504897] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 618.505689] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 9aab81df-f193-4ebc-a8ca-e4b8892b9597, please check neutron logs for more information. [ 618.505689] env[61906]: Removing descriptor: 17 [ 618.505689] env[61906]: ERROR nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9aab81df-f193-4ebc-a8ca-e4b8892b9597, please check neutron logs for more information. [ 618.505689] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Traceback (most recent call last): [ 618.505689] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 618.505689] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] yield resources [ 618.505689] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.505689] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] self.driver.spawn(context, instance, image_meta, [ 618.505689] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 618.505689] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.505689] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.505689] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] vm_ref = self.build_virtual_machine(instance, [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] for vif in network_info: [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] return self._sync_wrapper(fn, *args, **kwargs) [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] self.wait() [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] self[:] = self._gt.wait() [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] return self._exit_event.wait() [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.506079] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] result = hub.switch() [ 618.506524] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.506524] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] return self.greenlet.switch() [ 618.506524] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.506524] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] result = function(*args, **kwargs) [ 618.506524] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.506524] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] return func(*args, **kwargs) [ 618.506524] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.506524] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] raise e [ 618.506524] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.506524] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] nwinfo = self.network_api.allocate_for_instance( [ 618.506524] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.506524] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] created_port_ids = self._update_ports_for_instance( [ 618.506524] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.506856] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] with excutils.save_and_reraise_exception(): [ 618.506856] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.506856] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] self.force_reraise() [ 618.506856] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.506856] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] raise self.value [ 618.506856] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.506856] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] updated_port = self._update_port( [ 618.506856] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.506856] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] _ensure_no_port_binding_failure(port) [ 618.506856] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.506856] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] raise exception.PortBindingFailed(port_id=port['id']) [ 618.506856] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] nova.exception.PortBindingFailed: Binding failed for port 9aab81df-f193-4ebc-a8ca-e4b8892b9597, please check neutron logs for more information. [ 618.506856] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] [ 618.507200] env[61906]: INFO nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Terminating instance [ 618.509715] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "refresh_cache-b56dd330-d31c-4e42-baf9-d9897c55e633" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.551266] env[61906]: DEBUG nova.compute.utils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 618.557879] env[61906]: DEBUG nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 618.558078] env[61906]: DEBUG nova.network.neutron [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 618.627658] env[61906]: DEBUG nova.network.neutron [req-7464899d-3b68-429e-afbd-803c091aa8ab req-5664214f-d726-40d7-a840-b31b74e634ed service nova] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.632949] env[61906]: DEBUG nova.policy [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0c0b016b8f1a4c569d8a0e7749ba0a2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2d1343211154305bb672928d6506d94', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 618.841767] env[61906]: DEBUG nova.network.neutron [req-7464899d-3b68-429e-afbd-803c091aa8ab req-5664214f-d726-40d7-a840-b31b74e634ed service nova] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.059704] env[61906]: DEBUG nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 619.147103] env[61906]: DEBUG nova.network.neutron [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Successfully created port: 0e98277e-b125-4b5e-8318-627224b13767 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 619.353187] env[61906]: DEBUG oslo_concurrency.lockutils [req-7464899d-3b68-429e-afbd-803c091aa8ab req-5664214f-d726-40d7-a840-b31b74e634ed service nova] Releasing lock "refresh_cache-b56dd330-d31c-4e42-baf9-d9897c55e633" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.353734] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquired lock "refresh_cache-b56dd330-d31c-4e42-baf9-d9897c55e633" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.354046] env[61906]: DEBUG nova.network.neutron [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.559183] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4939f879-78c3-4fcd-b85a-7b984fd9dc6b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.573273] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9d3634-4682-47e0-8399-6ce532cf7aee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.607101] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5fc760-d1e4-4e48-a590-a6dd397170ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.617533] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e416007e-ef2c-447b-a409-6ec02cf20c47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.632073] env[61906]: DEBUG nova.compute.provider_tree [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.876705] env[61906]: DEBUG nova.network.neutron [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.010698] env[61906]: DEBUG nova.network.neutron [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.071208] env[61906]: DEBUG nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 620.110948] env[61906]: DEBUG nova.virt.hardware [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:05:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='35bcdaf3-fe7a-403b-a414-5aa03ee691bc',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-4916568',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 620.111184] env[61906]: DEBUG nova.virt.hardware [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 620.111578] env[61906]: DEBUG nova.virt.hardware [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 620.111578] env[61906]: DEBUG nova.virt.hardware [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 620.111681] env[61906]: DEBUG nova.virt.hardware [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 620.114568] env[61906]: DEBUG nova.virt.hardware [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 620.114568] env[61906]: DEBUG nova.virt.hardware [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 620.115297] env[61906]: DEBUG nova.virt.hardware [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 620.115931] env[61906]: DEBUG nova.virt.hardware [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 620.116237] env[61906]: DEBUG nova.virt.hardware [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 620.117281] env[61906]: DEBUG nova.virt.hardware [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 620.118210] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0991a8-63e5-4120-bb45-741b0c3b3558 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.128691] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce662a90-2d72-42b6-9c4e-a05fab386e7f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.135235] env[61906]: DEBUG nova.scheduler.client.report [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.217906] env[61906]: DEBUG nova.compute.manager [req-e1c78997-3af2-46ab-ba90-2a758d856c02 req-3012c0ae-1c57-4cbe-bda4-fe5a9cbcf0cb service nova] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Received event network-vif-deleted-9aab81df-f193-4ebc-a8ca-e4b8892b9597 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 620.513768] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Releasing lock "refresh_cache-b56dd330-d31c-4e42-baf9-d9897c55e633" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.514219] env[61906]: DEBUG nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 620.514448] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.514787] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e307e145-f7b8-4b18-80fe-12dafa76df80 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.524551] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90629f15-e752-45be-b47f-766bab81a5be {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.551793] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b56dd330-d31c-4e42-baf9-d9897c55e633 could not be found. [ 620.552078] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 620.552316] env[61906]: INFO nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Took 0.04 seconds to destroy the instance on the hypervisor. [ 620.552575] env[61906]: DEBUG oslo.service.loopingcall [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.552801] env[61906]: DEBUG nova.compute.manager [-] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.552887] env[61906]: DEBUG nova.network.neutron [-] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.568513] env[61906]: DEBUG nova.network.neutron [-] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.653063] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.606s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.653063] env[61906]: DEBUG nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 620.654098] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.569s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.687474] env[61906]: ERROR nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0e98277e-b125-4b5e-8318-627224b13767, please check neutron logs for more information. [ 620.687474] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 620.687474] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.687474] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 620.687474] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.687474] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 620.687474] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.687474] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 620.687474] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.687474] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 620.687474] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.687474] env[61906]: ERROR nova.compute.manager raise self.value [ 620.687474] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.687474] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 620.687474] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.687474] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 620.687945] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.687945] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 620.687945] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0e98277e-b125-4b5e-8318-627224b13767, please check neutron logs for more information. [ 620.687945] env[61906]: ERROR nova.compute.manager [ 620.687945] env[61906]: Traceback (most recent call last): [ 620.687945] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 620.687945] env[61906]: listener.cb(fileno) [ 620.687945] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.687945] env[61906]: result = function(*args, **kwargs) [ 620.687945] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 620.687945] env[61906]: return func(*args, **kwargs) [ 620.687945] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.687945] env[61906]: raise e [ 620.687945] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.687945] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 620.687945] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.687945] env[61906]: created_port_ids = self._update_ports_for_instance( [ 620.687945] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.687945] env[61906]: with excutils.save_and_reraise_exception(): [ 620.687945] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.687945] env[61906]: self.force_reraise() [ 620.687945] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.687945] env[61906]: raise self.value [ 620.687945] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.687945] env[61906]: updated_port = self._update_port( [ 620.687945] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.687945] env[61906]: _ensure_no_port_binding_failure(port) [ 620.687945] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.687945] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 620.688689] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 0e98277e-b125-4b5e-8318-627224b13767, please check neutron logs for more information. [ 620.688689] env[61906]: Removing descriptor: 17 [ 620.688689] env[61906]: ERROR nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0e98277e-b125-4b5e-8318-627224b13767, please check neutron logs for more information. [ 620.688689] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Traceback (most recent call last): [ 620.688689] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 620.688689] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] yield resources [ 620.688689] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 620.688689] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] self.driver.spawn(context, instance, image_meta, [ 620.688689] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 620.688689] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.688689] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.688689] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] vm_ref = self.build_virtual_machine(instance, [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] for vif in network_info: [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] return self._sync_wrapper(fn, *args, **kwargs) [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] self.wait() [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] self[:] = self._gt.wait() [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] return self._exit_event.wait() [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.689013] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] result = hub.switch() [ 620.689358] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.689358] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] return self.greenlet.switch() [ 620.689358] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.689358] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] result = function(*args, **kwargs) [ 620.689358] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 620.689358] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] return func(*args, **kwargs) [ 620.689358] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.689358] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] raise e [ 620.689358] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.689358] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] nwinfo = self.network_api.allocate_for_instance( [ 620.689358] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.689358] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] created_port_ids = self._update_ports_for_instance( [ 620.689358] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.689668] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] with excutils.save_and_reraise_exception(): [ 620.689668] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.689668] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] self.force_reraise() [ 620.689668] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.689668] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] raise self.value [ 620.689668] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.689668] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] updated_port = self._update_port( [ 620.689668] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.689668] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] _ensure_no_port_binding_failure(port) [ 620.689668] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.689668] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] raise exception.PortBindingFailed(port_id=port['id']) [ 620.689668] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] nova.exception.PortBindingFailed: Binding failed for port 0e98277e-b125-4b5e-8318-627224b13767, please check neutron logs for more information. [ 620.689668] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] [ 620.689987] env[61906]: INFO nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Terminating instance [ 620.693337] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquiring lock "refresh_cache-f9edb738-210e-4f85-8062-7759fa9033ea" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.693337] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquired lock "refresh_cache-f9edb738-210e-4f85-8062-7759fa9033ea" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.693337] env[61906]: DEBUG nova.network.neutron [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.073557] env[61906]: DEBUG nova.network.neutron [-] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.160791] env[61906]: DEBUG nova.compute.utils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.167025] env[61906]: DEBUG nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 621.167025] env[61906]: DEBUG nova.network.neutron [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 621.215051] env[61906]: DEBUG nova.network.neutron [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.244645] env[61906]: DEBUG nova.policy [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4763c3a0fede4c44ba578972d8d90601', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3554954fb95f4cf3876237804bead670', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 621.302132] env[61906]: DEBUG nova.network.neutron [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.582020] env[61906]: INFO nova.compute.manager [-] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Took 1.02 seconds to deallocate network for instance. [ 621.582020] env[61906]: DEBUG nova.compute.claims [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 621.582020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.637277] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6635aa-d86d-478a-9f9e-b1a6b1d4c3d3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.648903] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02b1f8f-8f63-4160-ab5b-290812dd2882 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.685098] env[61906]: DEBUG nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 621.689583] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b0fbfd-9d14-42d6-a2e9-d1a93c10b0cc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.703490] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a5cef9-616a-4bb3-9d39-7c8064bfb950 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.718241] env[61906]: DEBUG nova.compute.provider_tree [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.746516] env[61906]: DEBUG nova.network.neutron [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Successfully created port: af963971-9ac5-40d3-930b-b53aea6dc233 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 621.807194] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Releasing lock "refresh_cache-f9edb738-210e-4f85-8062-7759fa9033ea" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.807194] env[61906]: DEBUG nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 621.807194] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 621.807194] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0e240cb3-91b2-42fb-8715-004d1215ee57 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.814917] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49426daf-c2a0-4dac-9c93-a76497a6f0ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.840801] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f9edb738-210e-4f85-8062-7759fa9033ea could not be found. [ 621.841310] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 621.841702] env[61906]: INFO nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Took 0.04 seconds to destroy the instance on the hypervisor. [ 621.842120] env[61906]: DEBUG oslo.service.loopingcall [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.842536] env[61906]: DEBUG nova.compute.manager [-] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.842760] env[61906]: DEBUG nova.network.neutron [-] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 621.870636] env[61906]: DEBUG nova.network.neutron [-] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.225665] env[61906]: DEBUG nova.scheduler.client.report [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.270719] env[61906]: DEBUG nova.compute.manager [req-5d0af939-0d12-4839-9471-54fa1f075205 req-552ddf56-8783-4480-bdba-3260a73985f4 service nova] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Received event network-changed-0e98277e-b125-4b5e-8318-627224b13767 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 622.270917] env[61906]: DEBUG nova.compute.manager [req-5d0af939-0d12-4839-9471-54fa1f075205 req-552ddf56-8783-4480-bdba-3260a73985f4 service nova] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Refreshing instance network info cache due to event network-changed-0e98277e-b125-4b5e-8318-627224b13767. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 622.271172] env[61906]: DEBUG oslo_concurrency.lockutils [req-5d0af939-0d12-4839-9471-54fa1f075205 req-552ddf56-8783-4480-bdba-3260a73985f4 service nova] Acquiring lock "refresh_cache-f9edb738-210e-4f85-8062-7759fa9033ea" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.271340] env[61906]: DEBUG oslo_concurrency.lockutils [req-5d0af939-0d12-4839-9471-54fa1f075205 req-552ddf56-8783-4480-bdba-3260a73985f4 service nova] Acquired lock "refresh_cache-f9edb738-210e-4f85-8062-7759fa9033ea" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.271507] env[61906]: DEBUG nova.network.neutron [req-5d0af939-0d12-4839-9471-54fa1f075205 req-552ddf56-8783-4480-bdba-3260a73985f4 service nova] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Refreshing network info cache for port 0e98277e-b125-4b5e-8318-627224b13767 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 622.372121] env[61906]: DEBUG nova.network.neutron [-] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.699866] env[61906]: DEBUG nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 622.731865] env[61906]: DEBUG nova.virt.hardware [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 622.733279] env[61906]: DEBUG nova.virt.hardware [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 622.733279] env[61906]: DEBUG nova.virt.hardware [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.733279] env[61906]: DEBUG nova.virt.hardware [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 622.733279] env[61906]: DEBUG nova.virt.hardware [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.733279] env[61906]: DEBUG nova.virt.hardware [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 622.733451] env[61906]: DEBUG nova.virt.hardware [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 622.733451] env[61906]: DEBUG nova.virt.hardware [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 622.733451] env[61906]: DEBUG nova.virt.hardware [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 622.733533] env[61906]: DEBUG nova.virt.hardware [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 622.733667] env[61906]: DEBUG nova.virt.hardware [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 622.734504] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.080s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.735146] env[61906]: ERROR nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1879d0be-ab99-41ea-a229-33aba99ee775, please check neutron logs for more information. [ 622.735146] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Traceback (most recent call last): [ 622.735146] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 622.735146] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] self.driver.spawn(context, instance, image_meta, [ 622.735146] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 622.735146] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.735146] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.735146] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] vm_ref = self.build_virtual_machine(instance, [ 622.735146] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.735146] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.735146] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.735477] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] for vif in network_info: [ 622.735477] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.735477] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] return self._sync_wrapper(fn, *args, **kwargs) [ 622.735477] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.735477] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] self.wait() [ 622.735477] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.735477] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] self[:] = self._gt.wait() [ 622.735477] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.735477] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] return self._exit_event.wait() [ 622.735477] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 622.735477] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] result = hub.switch() [ 622.735477] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 622.735477] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] return self.greenlet.switch() [ 622.735835] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.735835] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] result = function(*args, **kwargs) [ 622.735835] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 622.735835] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] return func(*args, **kwargs) [ 622.735835] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.735835] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] raise e [ 622.735835] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.735835] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] nwinfo = self.network_api.allocate_for_instance( [ 622.735835] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.735835] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] created_port_ids = self._update_ports_for_instance( [ 622.735835] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.735835] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] with excutils.save_and_reraise_exception(): [ 622.735835] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.736208] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] self.force_reraise() [ 622.736208] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.736208] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] raise self.value [ 622.736208] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.736208] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] updated_port = self._update_port( [ 622.736208] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.736208] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] _ensure_no_port_binding_failure(port) [ 622.736208] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.736208] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] raise exception.PortBindingFailed(port_id=port['id']) [ 622.736208] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] nova.exception.PortBindingFailed: Binding failed for port 1879d0be-ab99-41ea-a229-33aba99ee775, please check neutron logs for more information. [ 622.736208] env[61906]: ERROR nova.compute.manager [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] [ 622.736521] env[61906]: DEBUG nova.compute.utils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Binding failed for port 1879d0be-ab99-41ea-a229-33aba99ee775, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 622.739975] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a38ed59-074b-4438-8fd6-0bf745340c77 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.740782] env[61906]: DEBUG nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Build of instance 40f3efa6-699f-4779-b999-c86435e04c0a was re-scheduled: Binding failed for port 1879d0be-ab99-41ea-a229-33aba99ee775, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 622.741262] env[61906]: DEBUG nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 622.742160] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "refresh_cache-40f3efa6-699f-4779-b999-c86435e04c0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.742160] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquired lock "refresh_cache-40f3efa6-699f-4779-b999-c86435e04c0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.742160] env[61906]: DEBUG nova.network.neutron [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 622.742848] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.333s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.744603] env[61906]: INFO nova.compute.claims [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.754040] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc98e761-018d-4653-b7f1-a006bffb3de9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.799233] env[61906]: DEBUG nova.network.neutron [req-5d0af939-0d12-4839-9471-54fa1f075205 req-552ddf56-8783-4480-bdba-3260a73985f4 service nova] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.875955] env[61906]: INFO nova.compute.manager [-] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Took 1.03 seconds to deallocate network for instance. [ 622.878438] env[61906]: DEBUG nova.compute.claims [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 622.882349] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.884212] env[61906]: DEBUG nova.network.neutron [req-5d0af939-0d12-4839-9471-54fa1f075205 req-552ddf56-8783-4480-bdba-3260a73985f4 service nova] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.272038] env[61906]: DEBUG nova.network.neutron [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.386488] env[61906]: DEBUG oslo_concurrency.lockutils [req-5d0af939-0d12-4839-9471-54fa1f075205 req-552ddf56-8783-4480-bdba-3260a73985f4 service nova] Releasing lock "refresh_cache-f9edb738-210e-4f85-8062-7759fa9033ea" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.386777] env[61906]: DEBUG nova.compute.manager [req-5d0af939-0d12-4839-9471-54fa1f075205 req-552ddf56-8783-4480-bdba-3260a73985f4 service nova] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Received event network-vif-deleted-0e98277e-b125-4b5e-8318-627224b13767 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 623.402463] env[61906]: DEBUG nova.network.neutron [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.425153] env[61906]: ERROR nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port af963971-9ac5-40d3-930b-b53aea6dc233, please check neutron logs for more information. [ 623.425153] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 623.425153] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.425153] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 623.425153] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.425153] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 623.425153] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.425153] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 623.425153] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.425153] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 623.425153] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.425153] env[61906]: ERROR nova.compute.manager raise self.value [ 623.425153] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.425153] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 623.425153] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.425153] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 623.425718] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.425718] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 623.425718] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port af963971-9ac5-40d3-930b-b53aea6dc233, please check neutron logs for more information. [ 623.425718] env[61906]: ERROR nova.compute.manager [ 623.425718] env[61906]: Traceback (most recent call last): [ 623.425718] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 623.425718] env[61906]: listener.cb(fileno) [ 623.425718] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.425718] env[61906]: result = function(*args, **kwargs) [ 623.425718] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 623.425718] env[61906]: return func(*args, **kwargs) [ 623.425718] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.425718] env[61906]: raise e [ 623.425718] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.425718] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 623.425718] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.425718] env[61906]: created_port_ids = self._update_ports_for_instance( [ 623.425718] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.425718] env[61906]: with excutils.save_and_reraise_exception(): [ 623.425718] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.425718] env[61906]: self.force_reraise() [ 623.425718] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.425718] env[61906]: raise self.value [ 623.425718] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.425718] env[61906]: updated_port = self._update_port( [ 623.425718] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.425718] env[61906]: _ensure_no_port_binding_failure(port) [ 623.425718] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.425718] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 623.426687] env[61906]: nova.exception.PortBindingFailed: Binding failed for port af963971-9ac5-40d3-930b-b53aea6dc233, please check neutron logs for more information. [ 623.426687] env[61906]: Removing descriptor: 17 [ 623.426687] env[61906]: ERROR nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port af963971-9ac5-40d3-930b-b53aea6dc233, please check neutron logs for more information. [ 623.426687] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Traceback (most recent call last): [ 623.426687] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 623.426687] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] yield resources [ 623.426687] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.426687] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] self.driver.spawn(context, instance, image_meta, [ 623.426687] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 623.426687] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.426687] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.426687] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] vm_ref = self.build_virtual_machine(instance, [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] for vif in network_info: [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] return self._sync_wrapper(fn, *args, **kwargs) [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] self.wait() [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] self[:] = self._gt.wait() [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] return self._exit_event.wait() [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.427048] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] result = hub.switch() [ 623.427450] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.427450] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] return self.greenlet.switch() [ 623.427450] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.427450] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] result = function(*args, **kwargs) [ 623.427450] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 623.427450] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] return func(*args, **kwargs) [ 623.427450] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.427450] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] raise e [ 623.427450] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.427450] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] nwinfo = self.network_api.allocate_for_instance( [ 623.427450] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.427450] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] created_port_ids = self._update_ports_for_instance( [ 623.427450] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.427810] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] with excutils.save_and_reraise_exception(): [ 623.427810] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.427810] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] self.force_reraise() [ 623.427810] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.427810] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] raise self.value [ 623.427810] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.427810] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] updated_port = self._update_port( [ 623.427810] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.427810] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] _ensure_no_port_binding_failure(port) [ 623.427810] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.427810] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] raise exception.PortBindingFailed(port_id=port['id']) [ 623.427810] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] nova.exception.PortBindingFailed: Binding failed for port af963971-9ac5-40d3-930b-b53aea6dc233, please check neutron logs for more information. [ 623.427810] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] [ 623.428200] env[61906]: INFO nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Terminating instance [ 623.428632] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Acquiring lock "refresh_cache-3e6732aa-a878-4bde-9299-abd779f2a109" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.428820] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Acquired lock "refresh_cache-3e6732aa-a878-4bde-9299-abd779f2a109" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.428989] env[61906]: DEBUG nova.network.neutron [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 623.907951] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Releasing lock "refresh_cache-40f3efa6-699f-4779-b999-c86435e04c0a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.907951] env[61906]: DEBUG nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 623.907951] env[61906]: DEBUG nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.907951] env[61906]: DEBUG nova.network.neutron [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 623.922712] env[61906]: DEBUG nova.network.neutron [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.956767] env[61906]: DEBUG nova.network.neutron [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.092232] env[61906]: DEBUG nova.network.neutron [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.152487] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4703e83d-bd65-424f-ae73-4fb1ef649c9e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.160555] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f45e1c-b7c2-4372-9f7b-76aa0f53bef7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.191977] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd6a5de-ec66-4d75-9f16-fcd495c8fa2a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.199526] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776e638a-da26-4fec-b73e-81df97e3f3c7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.213531] env[61906]: DEBUG nova.compute.provider_tree [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.414113] env[61906]: DEBUG nova.compute.manager [req-7af94ab9-0d36-4aa4-b762-9951179feb22 req-7a35a46c-2a5e-4c46-a4ce-379a408fee43 service nova] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Received event network-changed-af963971-9ac5-40d3-930b-b53aea6dc233 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 624.414113] env[61906]: DEBUG nova.compute.manager [req-7af94ab9-0d36-4aa4-b762-9951179feb22 req-7a35a46c-2a5e-4c46-a4ce-379a408fee43 service nova] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Refreshing instance network info cache due to event network-changed-af963971-9ac5-40d3-930b-b53aea6dc233. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 624.414195] env[61906]: DEBUG oslo_concurrency.lockutils [req-7af94ab9-0d36-4aa4-b762-9951179feb22 req-7a35a46c-2a5e-4c46-a4ce-379a408fee43 service nova] Acquiring lock "refresh_cache-3e6732aa-a878-4bde-9299-abd779f2a109" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.425719] env[61906]: DEBUG nova.network.neutron [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.596915] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Releasing lock "refresh_cache-3e6732aa-a878-4bde-9299-abd779f2a109" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.597604] env[61906]: DEBUG nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 624.597819] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 624.598158] env[61906]: DEBUG oslo_concurrency.lockutils [req-7af94ab9-0d36-4aa4-b762-9951179feb22 req-7a35a46c-2a5e-4c46-a4ce-379a408fee43 service nova] Acquired lock "refresh_cache-3e6732aa-a878-4bde-9299-abd779f2a109" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.598333] env[61906]: DEBUG nova.network.neutron [req-7af94ab9-0d36-4aa4-b762-9951179feb22 req-7a35a46c-2a5e-4c46-a4ce-379a408fee43 service nova] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Refreshing network info cache for port af963971-9ac5-40d3-930b-b53aea6dc233 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 624.599400] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b00d4c87-c888-4a69-97d1-d025677a0431 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.609190] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a6e407-8b09-4bb6-bd75-fb5ef5bda305 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.631022] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3e6732aa-a878-4bde-9299-abd779f2a109 could not be found. [ 624.631022] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 624.631022] env[61906]: INFO nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Took 0.03 seconds to destroy the instance on the hypervisor. [ 624.631022] env[61906]: DEBUG oslo.service.loopingcall [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.631216] env[61906]: DEBUG nova.compute.manager [-] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.631317] env[61906]: DEBUG nova.network.neutron [-] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 624.647038] env[61906]: DEBUG nova.network.neutron [-] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.717019] env[61906]: DEBUG nova.scheduler.client.report [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.880329] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Acquiring lock "e0ea5eb3-82cf-44f1-91ee-2c2861aefee4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.880561] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Lock "e0ea5eb3-82cf-44f1-91ee-2c2861aefee4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.928500] env[61906]: INFO nova.compute.manager [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 40f3efa6-699f-4779-b999-c86435e04c0a] Took 1.02 seconds to deallocate network for instance. [ 625.124628] env[61906]: DEBUG nova.network.neutron [req-7af94ab9-0d36-4aa4-b762-9951179feb22 req-7a35a46c-2a5e-4c46-a4ce-379a408fee43 service nova] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.148787] env[61906]: DEBUG nova.network.neutron [-] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.222219] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.222590] env[61906]: DEBUG nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 625.225810] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.334s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.229733] env[61906]: DEBUG nova.network.neutron [req-7af94ab9-0d36-4aa4-b762-9951179feb22 req-7a35a46c-2a5e-4c46-a4ce-379a408fee43 service nova] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.651825] env[61906]: INFO nova.compute.manager [-] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Took 1.02 seconds to deallocate network for instance. [ 625.654476] env[61906]: DEBUG nova.compute.claims [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 625.654717] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.733427] env[61906]: DEBUG nova.compute.utils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 625.734978] env[61906]: DEBUG oslo_concurrency.lockutils [req-7af94ab9-0d36-4aa4-b762-9951179feb22 req-7a35a46c-2a5e-4c46-a4ce-379a408fee43 service nova] Releasing lock "refresh_cache-3e6732aa-a878-4bde-9299-abd779f2a109" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.735227] env[61906]: DEBUG nova.compute.manager [req-7af94ab9-0d36-4aa4-b762-9951179feb22 req-7a35a46c-2a5e-4c46-a4ce-379a408fee43 service nova] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Received event network-vif-deleted-af963971-9ac5-40d3-930b-b53aea6dc233 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 625.735569] env[61906]: DEBUG nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 625.735727] env[61906]: DEBUG nova.network.neutron [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 625.774227] env[61906]: DEBUG nova.policy [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '43155dc89cd749c9862840ab2ea17bc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5eb8e11c074f4f4abfa3080006e300bc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 625.955267] env[61906]: INFO nova.scheduler.client.report [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Deleted allocations for instance 40f3efa6-699f-4779-b999-c86435e04c0a [ 626.069416] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284351e4-acc5-4a0f-8f00-246f384956e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.077537] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-298747f2-29cd-46b8-9bb3-119608edffcd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.107808] env[61906]: DEBUG nova.network.neutron [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Successfully created port: def891b7-a3b4-4b8e-bdbe-9e50e2f817c3 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.110164] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c519a915-50e3-4b12-8744-29634df1a5e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.117410] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef97fe58-f83d-4039-9a31-6f9c8cec6ccb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.130486] env[61906]: DEBUG nova.compute.provider_tree [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.239566] env[61906]: DEBUG nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 626.467399] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbe2fa8-1542-4d98-90a7-0963f19d1bf7 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "40f3efa6-699f-4779-b999-c86435e04c0a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.788s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.633130] env[61906]: DEBUG nova.scheduler.client.report [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.744172] env[61906]: INFO nova.virt.block_device [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Booting with volume d4fe3ed0-6de6-42d8-ac65-6a324ac1b224 at /dev/sda [ 626.793364] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d5e708f-af2f-40c4-be6c-92f0597c9d00 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.804454] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4774525e-b8ef-464a-92aa-18932a5018a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.826499] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-101222bf-fdb4-4df2-9a98-3c8a96f1dd9e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.834060] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c20447-eaf1-494f-952e-1d65ca6171ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.855017] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f83cc6-f2bd-4a5f-ba0c-14b896d50df2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.861278] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8fa72d2-821d-482c-a519-0679361695e2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.878182] env[61906]: DEBUG nova.virt.block_device [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Updating existing volume attachment record: 113e98b7-4acb-4f20-b3f5-2ad6cc3ae644 {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 626.908635] env[61906]: DEBUG nova.compute.manager [req-d365b04e-2c42-418f-8c66-16faf071476e req-8b1f87b5-6cdd-4a40-8ca1-42928c680c1b service nova] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Received event network-changed-def891b7-a3b4-4b8e-bdbe-9e50e2f817c3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 626.908904] env[61906]: DEBUG nova.compute.manager [req-d365b04e-2c42-418f-8c66-16faf071476e req-8b1f87b5-6cdd-4a40-8ca1-42928c680c1b service nova] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Refreshing instance network info cache due to event network-changed-def891b7-a3b4-4b8e-bdbe-9e50e2f817c3. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 626.909188] env[61906]: DEBUG oslo_concurrency.lockutils [req-d365b04e-2c42-418f-8c66-16faf071476e req-8b1f87b5-6cdd-4a40-8ca1-42928c680c1b service nova] Acquiring lock "refresh_cache-73a96c0e-a08b-46c9-b700-25544e17c73d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.909338] env[61906]: DEBUG oslo_concurrency.lockutils [req-d365b04e-2c42-418f-8c66-16faf071476e req-8b1f87b5-6cdd-4a40-8ca1-42928c680c1b service nova] Acquired lock "refresh_cache-73a96c0e-a08b-46c9-b700-25544e17c73d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.909507] env[61906]: DEBUG nova.network.neutron [req-d365b04e-2c42-418f-8c66-16faf071476e req-8b1f87b5-6cdd-4a40-8ca1-42928c680c1b service nova] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Refreshing network info cache for port def891b7-a3b4-4b8e-bdbe-9e50e2f817c3 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 626.969712] env[61906]: DEBUG nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 627.052711] env[61906]: ERROR nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port def891b7-a3b4-4b8e-bdbe-9e50e2f817c3, please check neutron logs for more information. [ 627.052711] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 627.052711] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.052711] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 627.052711] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.052711] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 627.052711] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.052711] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 627.052711] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.052711] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 627.052711] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.052711] env[61906]: ERROR nova.compute.manager raise self.value [ 627.052711] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.052711] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 627.052711] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.052711] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 627.053163] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.053163] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 627.053163] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port def891b7-a3b4-4b8e-bdbe-9e50e2f817c3, please check neutron logs for more information. [ 627.053163] env[61906]: ERROR nova.compute.manager [ 627.053163] env[61906]: Traceback (most recent call last): [ 627.053163] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 627.053163] env[61906]: listener.cb(fileno) [ 627.053163] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.053163] env[61906]: result = function(*args, **kwargs) [ 627.053163] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 627.053163] env[61906]: return func(*args, **kwargs) [ 627.053163] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.053163] env[61906]: raise e [ 627.053163] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.053163] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 627.053163] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.053163] env[61906]: created_port_ids = self._update_ports_for_instance( [ 627.053163] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.053163] env[61906]: with excutils.save_and_reraise_exception(): [ 627.053163] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.053163] env[61906]: self.force_reraise() [ 627.053163] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.053163] env[61906]: raise self.value [ 627.053163] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.053163] env[61906]: updated_port = self._update_port( [ 627.053163] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.053163] env[61906]: _ensure_no_port_binding_failure(port) [ 627.053163] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.053163] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 627.054030] env[61906]: nova.exception.PortBindingFailed: Binding failed for port def891b7-a3b4-4b8e-bdbe-9e50e2f817c3, please check neutron logs for more information. [ 627.054030] env[61906]: Removing descriptor: 17 [ 627.138608] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.913s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.139277] env[61906]: ERROR nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d85e5455-62ad-4a39-bb2f-c0f60076528c, please check neutron logs for more information. [ 627.139277] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Traceback (most recent call last): [ 627.139277] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.139277] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] self.driver.spawn(context, instance, image_meta, [ 627.139277] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 627.139277] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.139277] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.139277] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] vm_ref = self.build_virtual_machine(instance, [ 627.139277] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.139277] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.139277] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.139691] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] for vif in network_info: [ 627.139691] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.139691] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] return self._sync_wrapper(fn, *args, **kwargs) [ 627.139691] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.139691] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] self.wait() [ 627.139691] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.139691] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] self[:] = self._gt.wait() [ 627.139691] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.139691] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] return self._exit_event.wait() [ 627.139691] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.139691] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] result = hub.switch() [ 627.139691] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.139691] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] return self.greenlet.switch() [ 627.140289] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.140289] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] result = function(*args, **kwargs) [ 627.140289] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 627.140289] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] return func(*args, **kwargs) [ 627.140289] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.140289] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] raise e [ 627.140289] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.140289] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] nwinfo = self.network_api.allocate_for_instance( [ 627.140289] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.140289] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] created_port_ids = self._update_ports_for_instance( [ 627.140289] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.140289] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] with excutils.save_and_reraise_exception(): [ 627.140289] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.140676] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] self.force_reraise() [ 627.140676] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.140676] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] raise self.value [ 627.140676] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.140676] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] updated_port = self._update_port( [ 627.140676] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.140676] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] _ensure_no_port_binding_failure(port) [ 627.140676] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.140676] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] raise exception.PortBindingFailed(port_id=port['id']) [ 627.140676] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] nova.exception.PortBindingFailed: Binding failed for port d85e5455-62ad-4a39-bb2f-c0f60076528c, please check neutron logs for more information. [ 627.140676] env[61906]: ERROR nova.compute.manager [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] [ 627.141034] env[61906]: DEBUG nova.compute.utils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Binding failed for port d85e5455-62ad-4a39-bb2f-c0f60076528c, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 627.141170] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.202s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.141361] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.141516] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 627.141788] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.147s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.145334] env[61906]: INFO nova.compute.claims [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.148367] env[61906]: DEBUG nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Build of instance 081d06b3-c2b6-42f2-94b9-925493590e55 was re-scheduled: Binding failed for port d85e5455-62ad-4a39-bb2f-c0f60076528c, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 627.148685] env[61906]: DEBUG nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 627.148903] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "refresh_cache-081d06b3-c2b6-42f2-94b9-925493590e55" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.149100] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquired lock "refresh_cache-081d06b3-c2b6-42f2-94b9-925493590e55" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.149270] env[61906]: DEBUG nova.network.neutron [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.150853] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d539472-c7a7-4fcf-87bf-dd5af26ccdb8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.161049] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd1f892-1eeb-495c-8656-de06994eeffc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.176457] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c85b8a5-afb4-4c16-94f4-c0adeb0390b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.184286] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a725f190-dad4-4883-82f7-06727337ac71 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.212084] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181511MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 627.212202] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.428508] env[61906]: DEBUG nova.network.neutron [req-d365b04e-2c42-418f-8c66-16faf071476e req-8b1f87b5-6cdd-4a40-8ca1-42928c680c1b service nova] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.494931] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.525322] env[61906]: DEBUG nova.network.neutron [req-d365b04e-2c42-418f-8c66-16faf071476e req-8b1f87b5-6cdd-4a40-8ca1-42928c680c1b service nova] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.678264] env[61906]: DEBUG nova.network.neutron [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.754842] env[61906]: DEBUG nova.network.neutron [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.028387] env[61906]: DEBUG oslo_concurrency.lockutils [req-d365b04e-2c42-418f-8c66-16faf071476e req-8b1f87b5-6cdd-4a40-8ca1-42928c680c1b service nova] Releasing lock "refresh_cache-73a96c0e-a08b-46c9-b700-25544e17c73d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.257041] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Releasing lock "refresh_cache-081d06b3-c2b6-42f2-94b9-925493590e55" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.257384] env[61906]: DEBUG nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 628.257556] env[61906]: DEBUG nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.257709] env[61906]: DEBUG nova.network.neutron [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.274660] env[61906]: DEBUG nova.network.neutron [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.459401] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a046b4-142a-42ef-b439-288852ae22bc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.467049] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9213f43-d63c-408b-82ba-c39ee08db28b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.495519] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-344bc84a-6f20-47dd-b50e-555b29b2c200 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.502328] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98512a7-f640-4597-bfb7-0942cc3e8262 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.514866] env[61906]: DEBUG nova.compute.provider_tree [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.778248] env[61906]: DEBUG nova.network.neutron [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.933144] env[61906]: DEBUG nova.compute.manager [req-408d0c04-ef53-412a-9962-ca3e5ee10ce8 req-e929f79e-1bcc-4a3f-8b2b-0000abb76252 service nova] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Received event network-vif-deleted-def891b7-a3b4-4b8e-bdbe-9e50e2f817c3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.018110] env[61906]: DEBUG nova.scheduler.client.report [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.063036] env[61906]: DEBUG nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 629.063685] env[61906]: DEBUG nova.virt.hardware [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 629.063911] env[61906]: DEBUG nova.virt.hardware [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 629.064079] env[61906]: DEBUG nova.virt.hardware [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.064265] env[61906]: DEBUG nova.virt.hardware [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 629.064409] env[61906]: DEBUG nova.virt.hardware [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.064552] env[61906]: DEBUG nova.virt.hardware [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 629.064754] env[61906]: DEBUG nova.virt.hardware [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 629.064909] env[61906]: DEBUG nova.virt.hardware [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 629.065082] env[61906]: DEBUG nova.virt.hardware [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 629.065245] env[61906]: DEBUG nova.virt.hardware [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 629.065419] env[61906]: DEBUG nova.virt.hardware [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.066551] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda982c2-2f2b-49fb-b765-c36807e1e158 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.076055] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9485d6-75a1-46a6-9836-116e56049677 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.091733] env[61906]: ERROR nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port def891b7-a3b4-4b8e-bdbe-9e50e2f817c3, please check neutron logs for more information. [ 629.091733] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Traceback (most recent call last): [ 629.091733] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 629.091733] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] yield resources [ 629.091733] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.091733] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] self.driver.spawn(context, instance, image_meta, [ 629.091733] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 629.091733] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.091733] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.091733] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] vm_ref = self.build_virtual_machine(instance, [ 629.091733] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.092166] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.092166] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.092166] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] for vif in network_info: [ 629.092166] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.092166] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] return self._sync_wrapper(fn, *args, **kwargs) [ 629.092166] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.092166] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] self.wait() [ 629.092166] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.092166] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] self[:] = self._gt.wait() [ 629.092166] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.092166] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] return self._exit_event.wait() [ 629.092166] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 629.092166] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] current.throw(*self._exc) [ 629.092546] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.092546] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] result = function(*args, **kwargs) [ 629.092546] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 629.092546] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] return func(*args, **kwargs) [ 629.092546] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.092546] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] raise e [ 629.092546] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.092546] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] nwinfo = self.network_api.allocate_for_instance( [ 629.092546] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.092546] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] created_port_ids = self._update_ports_for_instance( [ 629.092546] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.092546] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] with excutils.save_and_reraise_exception(): [ 629.092546] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.092928] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] self.force_reraise() [ 629.092928] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.092928] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] raise self.value [ 629.092928] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.092928] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] updated_port = self._update_port( [ 629.092928] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.092928] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] _ensure_no_port_binding_failure(port) [ 629.092928] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.092928] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] raise exception.PortBindingFailed(port_id=port['id']) [ 629.092928] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] nova.exception.PortBindingFailed: Binding failed for port def891b7-a3b4-4b8e-bdbe-9e50e2f817c3, please check neutron logs for more information. [ 629.092928] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] [ 629.092928] env[61906]: INFO nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Terminating instance [ 629.093997] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Acquiring lock "refresh_cache-73a96c0e-a08b-46c9-b700-25544e17c73d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.094168] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Acquired lock "refresh_cache-73a96c0e-a08b-46c9-b700-25544e17c73d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.094336] env[61906]: DEBUG nova.network.neutron [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 629.281703] env[61906]: INFO nova.compute.manager [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: 081d06b3-c2b6-42f2-94b9-925493590e55] Took 1.02 seconds to deallocate network for instance. [ 629.523459] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.524065] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 629.526888] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.936s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.620260] env[61906]: DEBUG nova.network.neutron [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.701869] env[61906]: DEBUG nova.network.neutron [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.034324] env[61906]: DEBUG nova.compute.utils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 630.036239] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 630.036404] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 630.076579] env[61906]: DEBUG nova.policy [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2eb35d1ddff445d976412663620d6f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '853dcd3ea145424580e469edfa56c9c5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 630.204510] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Releasing lock "refresh_cache-73a96c0e-a08b-46c9-b700-25544e17c73d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.205075] env[61906]: DEBUG nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 630.205552] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-13a6f908-0e86-4e40-81e4-7276aadf170b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.214482] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33438b84-476b-45b0-b8e7-cf1b270f5f8c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.238758] env[61906]: WARNING nova.virt.vmwareapi.driver [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 73a96c0e-a08b-46c9-b700-25544e17c73d could not be found. [ 630.238966] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 630.241265] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6503b489-5cbf-4313-bf06-dcc7177573d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.250265] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-681660b5-5381-45cf-814a-c5f29e024c0c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.275386] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 73a96c0e-a08b-46c9-b700-25544e17c73d could not be found. [ 630.275386] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 630.275619] env[61906]: INFO nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Took 0.07 seconds to destroy the instance on the hypervisor. [ 630.275828] env[61906]: DEBUG oslo.service.loopingcall [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 630.278435] env[61906]: DEBUG nova.compute.manager [-] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 630.278435] env[61906]: DEBUG nova.network.neutron [-] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.300543] env[61906]: DEBUG nova.network.neutron [-] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.309708] env[61906]: INFO nova.scheduler.client.report [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Deleted allocations for instance 081d06b3-c2b6-42f2-94b9-925493590e55 [ 630.389770] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Successfully created port: 116f6b3d-5d45-4f71-ac3d-df6f8361e722 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.403761] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189567af-ea5d-44f7-a6ea-cc5571ea8da1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.418025] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db661b1-dce3-432d-b39b-eec3c1fdd882 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.441441] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f5d044-cb0a-40fd-b508-6bf76ab5d3fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.448508] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90de8126-6ca5-4fd1-bf11-fed03d787b2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.461996] env[61906]: DEBUG nova.compute.provider_tree [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.539287] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 630.802608] env[61906]: DEBUG nova.network.neutron [-] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.825263] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a07ff4d4-5e10-4cbb-9dc6-10da7e24a3b9 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "081d06b3-c2b6-42f2-94b9-925493590e55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.206s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.966222] env[61906]: DEBUG nova.scheduler.client.report [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.162684] env[61906]: DEBUG nova.compute.manager [req-706efa72-dbb1-4e34-8a42-d5a2a8c7f060 req-6fe00937-7522-4593-ad80-c833dc2c13b1 service nova] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Received event network-changed-116f6b3d-5d45-4f71-ac3d-df6f8361e722 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.162923] env[61906]: DEBUG nova.compute.manager [req-706efa72-dbb1-4e34-8a42-d5a2a8c7f060 req-6fe00937-7522-4593-ad80-c833dc2c13b1 service nova] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Refreshing instance network info cache due to event network-changed-116f6b3d-5d45-4f71-ac3d-df6f8361e722. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 631.163129] env[61906]: DEBUG oslo_concurrency.lockutils [req-706efa72-dbb1-4e34-8a42-d5a2a8c7f060 req-6fe00937-7522-4593-ad80-c833dc2c13b1 service nova] Acquiring lock "refresh_cache-0c7d1af9-54e9-435d-a0cc-687f793d4f43" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.163346] env[61906]: DEBUG oslo_concurrency.lockutils [req-706efa72-dbb1-4e34-8a42-d5a2a8c7f060 req-6fe00937-7522-4593-ad80-c833dc2c13b1 service nova] Acquired lock "refresh_cache-0c7d1af9-54e9-435d-a0cc-687f793d4f43" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.163526] env[61906]: DEBUG nova.network.neutron [req-706efa72-dbb1-4e34-8a42-d5a2a8c7f060 req-6fe00937-7522-4593-ad80-c833dc2c13b1 service nova] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Refreshing network info cache for port 116f6b3d-5d45-4f71-ac3d-df6f8361e722 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 631.305106] env[61906]: INFO nova.compute.manager [-] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Took 1.03 seconds to deallocate network for instance. [ 631.327529] env[61906]: DEBUG nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 631.368109] env[61906]: ERROR nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 116f6b3d-5d45-4f71-ac3d-df6f8361e722, please check neutron logs for more information. [ 631.368109] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.368109] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.368109] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.368109] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.368109] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.368109] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.368109] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.368109] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.368109] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 631.368109] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.368109] env[61906]: ERROR nova.compute.manager raise self.value [ 631.368109] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.368109] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.368109] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.368109] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.368624] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.368624] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.368624] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 116f6b3d-5d45-4f71-ac3d-df6f8361e722, please check neutron logs for more information. [ 631.368624] env[61906]: ERROR nova.compute.manager [ 631.368624] env[61906]: Traceback (most recent call last): [ 631.368624] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.368624] env[61906]: listener.cb(fileno) [ 631.368624] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.368624] env[61906]: result = function(*args, **kwargs) [ 631.368624] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.368624] env[61906]: return func(*args, **kwargs) [ 631.368624] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.368624] env[61906]: raise e [ 631.368624] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.368624] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 631.368624] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.368624] env[61906]: created_port_ids = self._update_ports_for_instance( [ 631.368624] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.368624] env[61906]: with excutils.save_and_reraise_exception(): [ 631.368624] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.368624] env[61906]: self.force_reraise() [ 631.368624] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.368624] env[61906]: raise self.value [ 631.368624] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.368624] env[61906]: updated_port = self._update_port( [ 631.368624] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.368624] env[61906]: _ensure_no_port_binding_failure(port) [ 631.368624] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.368624] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.369460] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 116f6b3d-5d45-4f71-ac3d-df6f8361e722, please check neutron logs for more information. [ 631.369460] env[61906]: Removing descriptor: 17 [ 631.471399] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.944s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.472095] env[61906]: ERROR nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7df4399b-0ff5-4b9b-94b8-ecc1547ee394, please check neutron logs for more information. [ 631.472095] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Traceback (most recent call last): [ 631.472095] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.472095] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] self.driver.spawn(context, instance, image_meta, [ 631.472095] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 631.472095] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.472095] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.472095] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] vm_ref = self.build_virtual_machine(instance, [ 631.472095] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.472095] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.472095] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.472492] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] for vif in network_info: [ 631.472492] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.472492] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] return self._sync_wrapper(fn, *args, **kwargs) [ 631.472492] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.472492] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] self.wait() [ 631.472492] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.472492] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] self[:] = self._gt.wait() [ 631.472492] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.472492] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] return self._exit_event.wait() [ 631.472492] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.472492] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] result = hub.switch() [ 631.472492] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.472492] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] return self.greenlet.switch() [ 631.473013] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.473013] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] result = function(*args, **kwargs) [ 631.473013] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.473013] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] return func(*args, **kwargs) [ 631.473013] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.473013] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] raise e [ 631.473013] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.473013] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] nwinfo = self.network_api.allocate_for_instance( [ 631.473013] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.473013] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] created_port_ids = self._update_ports_for_instance( [ 631.473013] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.473013] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] with excutils.save_and_reraise_exception(): [ 631.473013] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.473753] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] self.force_reraise() [ 631.473753] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.473753] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] raise self.value [ 631.473753] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.473753] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] updated_port = self._update_port( [ 631.473753] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.473753] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] _ensure_no_port_binding_failure(port) [ 631.473753] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.473753] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] raise exception.PortBindingFailed(port_id=port['id']) [ 631.473753] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] nova.exception.PortBindingFailed: Binding failed for port 7df4399b-0ff5-4b9b-94b8-ecc1547ee394, please check neutron logs for more information. [ 631.473753] env[61906]: ERROR nova.compute.manager [instance: ebea667f-6495-41c0-88b3-a9c61390558d] [ 631.474204] env[61906]: DEBUG nova.compute.utils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Binding failed for port 7df4399b-0ff5-4b9b-94b8-ecc1547ee394, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 631.474204] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.786s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.475549] env[61906]: INFO nova.compute.claims [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.478244] env[61906]: DEBUG nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Build of instance ebea667f-6495-41c0-88b3-a9c61390558d was re-scheduled: Binding failed for port 7df4399b-0ff5-4b9b-94b8-ecc1547ee394, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 631.478683] env[61906]: DEBUG nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 631.478903] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Acquiring lock "refresh_cache-ebea667f-6495-41c0-88b3-a9c61390558d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.479069] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Acquired lock "refresh_cache-ebea667f-6495-41c0-88b3-a9c61390558d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.479220] env[61906]: DEBUG nova.network.neutron [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.548598] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 631.574146] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 631.574497] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 631.574663] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 631.574849] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 631.575405] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 631.575405] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 631.575405] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 631.575602] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 631.575689] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 631.575812] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 631.575982] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 631.576842] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d238a59-8a9f-420f-af06-8851b01409e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.584961] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de727ec9-15d4-4634-98c0-80afb0bfb8fd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.598177] env[61906]: ERROR nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 116f6b3d-5d45-4f71-ac3d-df6f8361e722, please check neutron logs for more information. [ 631.598177] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Traceback (most recent call last): [ 631.598177] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 631.598177] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] yield resources [ 631.598177] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.598177] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] self.driver.spawn(context, instance, image_meta, [ 631.598177] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 631.598177] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.598177] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.598177] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] vm_ref = self.build_virtual_machine(instance, [ 631.598177] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.598575] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.598575] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.598575] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] for vif in network_info: [ 631.598575] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.598575] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] return self._sync_wrapper(fn, *args, **kwargs) [ 631.598575] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.598575] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] self.wait() [ 631.598575] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.598575] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] self[:] = self._gt.wait() [ 631.598575] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.598575] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] return self._exit_event.wait() [ 631.598575] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 631.598575] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] current.throw(*self._exc) [ 631.598903] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.598903] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] result = function(*args, **kwargs) [ 631.598903] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.598903] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] return func(*args, **kwargs) [ 631.598903] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.598903] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] raise e [ 631.598903] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.598903] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] nwinfo = self.network_api.allocate_for_instance( [ 631.598903] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.598903] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] created_port_ids = self._update_ports_for_instance( [ 631.598903] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.598903] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] with excutils.save_and_reraise_exception(): [ 631.598903] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.599542] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] self.force_reraise() [ 631.599542] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.599542] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] raise self.value [ 631.599542] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.599542] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] updated_port = self._update_port( [ 631.599542] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.599542] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] _ensure_no_port_binding_failure(port) [ 631.599542] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.599542] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] raise exception.PortBindingFailed(port_id=port['id']) [ 631.599542] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] nova.exception.PortBindingFailed: Binding failed for port 116f6b3d-5d45-4f71-ac3d-df6f8361e722, please check neutron logs for more information. [ 631.599542] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] [ 631.599542] env[61906]: INFO nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Terminating instance [ 631.600764] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "refresh_cache-0c7d1af9-54e9-435d-a0cc-687f793d4f43" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.684982] env[61906]: DEBUG nova.network.neutron [req-706efa72-dbb1-4e34-8a42-d5a2a8c7f060 req-6fe00937-7522-4593-ad80-c833dc2c13b1 service nova] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.768211] env[61906]: DEBUG nova.network.neutron [req-706efa72-dbb1-4e34-8a42-d5a2a8c7f060 req-6fe00937-7522-4593-ad80-c833dc2c13b1 service nova] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.851147] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.860295] env[61906]: INFO nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Took 0.56 seconds to detach 1 volumes for instance. [ 631.862400] env[61906]: DEBUG nova.compute.claims [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 631.862578] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.000129] env[61906]: DEBUG nova.network.neutron [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.059579] env[61906]: DEBUG nova.network.neutron [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.270612] env[61906]: DEBUG oslo_concurrency.lockutils [req-706efa72-dbb1-4e34-8a42-d5a2a8c7f060 req-6fe00937-7522-4593-ad80-c833dc2c13b1 service nova] Releasing lock "refresh_cache-0c7d1af9-54e9-435d-a0cc-687f793d4f43" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.271098] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquired lock "refresh_cache-0c7d1af9-54e9-435d-a0cc-687f793d4f43" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.271334] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 632.562725] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Releasing lock "refresh_cache-ebea667f-6495-41c0-88b3-a9c61390558d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.562725] env[61906]: DEBUG nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 632.562725] env[61906]: DEBUG nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.562890] env[61906]: DEBUG nova.network.neutron [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.579798] env[61906]: DEBUG nova.network.neutron [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.782621] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43bb09d0-642a-4c51-a467-481062482c7b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.791347] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.793649] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c5f610-db6e-437f-aee6-4e96528273c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.824030] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195948e6-ad05-427f-9307-155506a1dd55 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.831021] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc002c5c-a095-48c6-80d6-a4e55d08f909 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.844077] env[61906]: DEBUG nova.compute.provider_tree [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.902616] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.082034] env[61906]: DEBUG nova.network.neutron [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.184746] env[61906]: DEBUG nova.compute.manager [req-363c9b93-1a83-42f9-86a5-944c58cabb45 req-f81fd12f-78e2-4099-b3e2-33e838b4b63b service nova] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Received event network-vif-deleted-116f6b3d-5d45-4f71-ac3d-df6f8361e722 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 633.348105] env[61906]: DEBUG nova.scheduler.client.report [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.405062] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Releasing lock "refresh_cache-0c7d1af9-54e9-435d-a0cc-687f793d4f43" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.405515] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 633.405740] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 633.406041] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-acdb5b0f-f48f-4c0b-8617-f012275a2a6a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.415063] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39edac9-05c2-4ace-8e64-9ee11267dd3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.437325] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0c7d1af9-54e9-435d-a0cc-687f793d4f43 could not be found. [ 633.437554] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 633.437730] env[61906]: INFO nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Took 0.03 seconds to destroy the instance on the hypervisor. [ 633.438151] env[61906]: DEBUG oslo.service.loopingcall [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.438213] env[61906]: DEBUG nova.compute.manager [-] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.438276] env[61906]: DEBUG nova.network.neutron [-] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 633.453264] env[61906]: DEBUG nova.network.neutron [-] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.585212] env[61906]: INFO nova.compute.manager [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] [instance: ebea667f-6495-41c0-88b3-a9c61390558d] Took 1.02 seconds to deallocate network for instance. [ 633.852937] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.853655] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 633.856531] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.134s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.858325] env[61906]: INFO nova.compute.claims [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.955729] env[61906]: DEBUG nova.network.neutron [-] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.362230] env[61906]: DEBUG nova.compute.utils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.365834] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 634.366077] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 634.412376] env[61906]: DEBUG nova.policy [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2eb35d1ddff445d976412663620d6f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '853dcd3ea145424580e469edfa56c9c5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 634.458260] env[61906]: INFO nova.compute.manager [-] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Took 1.02 seconds to deallocate network for instance. [ 634.460797] env[61906]: DEBUG nova.compute.claims [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 634.461061] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.617310] env[61906]: INFO nova.scheduler.client.report [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Deleted allocations for instance ebea667f-6495-41c0-88b3-a9c61390558d [ 634.719708] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Successfully created port: 7df09b9b-9912-4b71-9cd2-6e0183ee2bde {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.867329] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 635.124581] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ea5ac371-2bdd-4bb0-9b87-6d83f0c02d96 tempest-ServerDiagnosticsNegativeTest-1129709807 tempest-ServerDiagnosticsNegativeTest-1129709807-project-member] Lock "ebea667f-6495-41c0-88b3-a9c61390558d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.873s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.215720] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c2de98c-d43e-44b3-8a56-50d018dac7bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.223365] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bdb8a79-7dd0-475e-baa3-ab364bd9321f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.254312] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7decef89-78bc-46b3-a1c1-c8915703ed09 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.261880] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed1c650-0d78-469e-82b2-256b7998a21f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.275386] env[61906]: DEBUG nova.compute.provider_tree [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.532695] env[61906]: DEBUG nova.compute.manager [req-9312718c-f66b-486e-9114-8d1ca3aa7cd3 req-4638a6df-6c3b-445f-b855-9b55d9180f74 service nova] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Received event network-changed-7df09b9b-9912-4b71-9cd2-6e0183ee2bde {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 635.532963] env[61906]: DEBUG nova.compute.manager [req-9312718c-f66b-486e-9114-8d1ca3aa7cd3 req-4638a6df-6c3b-445f-b855-9b55d9180f74 service nova] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Refreshing instance network info cache due to event network-changed-7df09b9b-9912-4b71-9cd2-6e0183ee2bde. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 635.533129] env[61906]: DEBUG oslo_concurrency.lockutils [req-9312718c-f66b-486e-9114-8d1ca3aa7cd3 req-4638a6df-6c3b-445f-b855-9b55d9180f74 service nova] Acquiring lock "refresh_cache-7aee8304-8d07-4aa2-867a-2eeeeacbabc8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.533190] env[61906]: DEBUG oslo_concurrency.lockutils [req-9312718c-f66b-486e-9114-8d1ca3aa7cd3 req-4638a6df-6c3b-445f-b855-9b55d9180f74 service nova] Acquired lock "refresh_cache-7aee8304-8d07-4aa2-867a-2eeeeacbabc8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.533424] env[61906]: DEBUG nova.network.neutron [req-9312718c-f66b-486e-9114-8d1ca3aa7cd3 req-4638a6df-6c3b-445f-b855-9b55d9180f74 service nova] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Refreshing network info cache for port 7df09b9b-9912-4b71-9cd2-6e0183ee2bde {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 635.627386] env[61906]: DEBUG nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 635.723865] env[61906]: ERROR nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7df09b9b-9912-4b71-9cd2-6e0183ee2bde, please check neutron logs for more information. [ 635.723865] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.723865] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.723865] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.723865] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.723865] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.723865] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.723865] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.723865] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.723865] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 635.723865] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.723865] env[61906]: ERROR nova.compute.manager raise self.value [ 635.723865] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.723865] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.723865] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.723865] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.724715] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.724715] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.724715] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7df09b9b-9912-4b71-9cd2-6e0183ee2bde, please check neutron logs for more information. [ 635.724715] env[61906]: ERROR nova.compute.manager [ 635.724715] env[61906]: Traceback (most recent call last): [ 635.724715] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.724715] env[61906]: listener.cb(fileno) [ 635.724715] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.724715] env[61906]: result = function(*args, **kwargs) [ 635.724715] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 635.724715] env[61906]: return func(*args, **kwargs) [ 635.724715] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.724715] env[61906]: raise e [ 635.724715] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.724715] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 635.724715] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.724715] env[61906]: created_port_ids = self._update_ports_for_instance( [ 635.724715] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.724715] env[61906]: with excutils.save_and_reraise_exception(): [ 635.724715] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.724715] env[61906]: self.force_reraise() [ 635.724715] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.724715] env[61906]: raise self.value [ 635.724715] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.724715] env[61906]: updated_port = self._update_port( [ 635.724715] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.724715] env[61906]: _ensure_no_port_binding_failure(port) [ 635.724715] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.724715] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.725905] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 7df09b9b-9912-4b71-9cd2-6e0183ee2bde, please check neutron logs for more information. [ 635.725905] env[61906]: Removing descriptor: 17 [ 635.778677] env[61906]: DEBUG nova.scheduler.client.report [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.878611] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 635.909488] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 635.909737] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 635.909892] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.910090] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 635.910239] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.910460] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 635.910671] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 635.910875] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 635.910875] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 635.911058] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 635.911293] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.912167] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4640cf-f328-4086-a32c-dfed41d1d20b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.920483] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70498b95-15cf-4028-92f6-c0934e62ea6e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.935133] env[61906]: ERROR nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7df09b9b-9912-4b71-9cd2-6e0183ee2bde, please check neutron logs for more information. [ 635.935133] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Traceback (most recent call last): [ 635.935133] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 635.935133] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] yield resources [ 635.935133] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.935133] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] self.driver.spawn(context, instance, image_meta, [ 635.935133] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 635.935133] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.935133] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.935133] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] vm_ref = self.build_virtual_machine(instance, [ 635.935133] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.936278] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.936278] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.936278] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] for vif in network_info: [ 635.936278] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.936278] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] return self._sync_wrapper(fn, *args, **kwargs) [ 635.936278] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.936278] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] self.wait() [ 635.936278] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.936278] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] self[:] = self._gt.wait() [ 635.936278] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.936278] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] return self._exit_event.wait() [ 635.936278] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 635.936278] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] current.throw(*self._exc) [ 635.937112] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.937112] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] result = function(*args, **kwargs) [ 635.937112] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 635.937112] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] return func(*args, **kwargs) [ 635.937112] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.937112] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] raise e [ 635.937112] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.937112] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] nwinfo = self.network_api.allocate_for_instance( [ 635.937112] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.937112] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] created_port_ids = self._update_ports_for_instance( [ 635.937112] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.937112] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] with excutils.save_and_reraise_exception(): [ 635.937112] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.937681] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] self.force_reraise() [ 635.937681] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.937681] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] raise self.value [ 635.937681] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.937681] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] updated_port = self._update_port( [ 635.937681] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.937681] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] _ensure_no_port_binding_failure(port) [ 635.937681] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.937681] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] raise exception.PortBindingFailed(port_id=port['id']) [ 635.937681] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] nova.exception.PortBindingFailed: Binding failed for port 7df09b9b-9912-4b71-9cd2-6e0183ee2bde, please check neutron logs for more information. [ 635.937681] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] [ 635.937681] env[61906]: INFO nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Terminating instance [ 635.940728] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "refresh_cache-7aee8304-8d07-4aa2-867a-2eeeeacbabc8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.051985] env[61906]: DEBUG nova.network.neutron [req-9312718c-f66b-486e-9114-8d1ca3aa7cd3 req-4638a6df-6c3b-445f-b855-9b55d9180f74 service nova] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.158575] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.164083] env[61906]: DEBUG nova.network.neutron [req-9312718c-f66b-486e-9114-8d1ca3aa7cd3 req-4638a6df-6c3b-445f-b855-9b55d9180f74 service nova] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.284206] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.284765] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 636.289551] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.710s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.667271] env[61906]: DEBUG oslo_concurrency.lockutils [req-9312718c-f66b-486e-9114-8d1ca3aa7cd3 req-4638a6df-6c3b-445f-b855-9b55d9180f74 service nova] Releasing lock "refresh_cache-7aee8304-8d07-4aa2-867a-2eeeeacbabc8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.667728] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquired lock "refresh_cache-7aee8304-8d07-4aa2-867a-2eeeeacbabc8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.667920] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.797143] env[61906]: DEBUG nova.compute.utils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.801201] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 636.801201] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 636.840393] env[61906]: DEBUG nova.policy [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2eb35d1ddff445d976412663620d6f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '853dcd3ea145424580e469edfa56c9c5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 637.146327] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2738a5c3-0c06-46d0-a749-ab7fbd22214c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.155362] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec166f7-66dc-484c-8001-d0f347909c76 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.162279] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Successfully created port: efa75200-e931-43ca-b12e-4fb50fd45b58 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 637.190771] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16a97ec-6ee2-49d9-a1d4-16a9dd85ffb6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.198095] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb940f8-ac2e-4139-a204-ccc60f37e833 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.212650] env[61906]: DEBUG nova.compute.provider_tree [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.217377] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.300248] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 637.516257] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.559310] env[61906]: DEBUG nova.compute.manager [req-bdeaa426-c107-437c-a6e4-f4c7ecf07d71 req-98c80782-2491-44ed-b703-49f98d234302 service nova] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Received event network-vif-deleted-7df09b9b-9912-4b71-9cd2-6e0183ee2bde {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.720194] env[61906]: DEBUG nova.scheduler.client.report [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.019425] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Releasing lock "refresh_cache-7aee8304-8d07-4aa2-867a-2eeeeacbabc8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.019776] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 638.019974] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 638.020290] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34e406c3-62a9-4a2f-9be2-9d06063aa7f9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.029603] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e011ef27-49eb-4dbc-b27a-14f6b80ce4e0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.051695] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7aee8304-8d07-4aa2-867a-2eeeeacbabc8 could not be found. [ 638.052013] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 638.052103] env[61906]: INFO nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Took 0.03 seconds to destroy the instance on the hypervisor. [ 638.052351] env[61906]: DEBUG oslo.service.loopingcall [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 638.053239] env[61906]: DEBUG nova.compute.manager [-] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.053239] env[61906]: DEBUG nova.network.neutron [-] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 638.079807] env[61906]: DEBUG nova.network.neutron [-] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.157856] env[61906]: ERROR nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port efa75200-e931-43ca-b12e-4fb50fd45b58, please check neutron logs for more information. [ 638.157856] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 638.157856] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.157856] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 638.157856] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.157856] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 638.157856] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.157856] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 638.157856] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.157856] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 638.157856] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.157856] env[61906]: ERROR nova.compute.manager raise self.value [ 638.157856] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.157856] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 638.157856] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.157856] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 638.158650] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.158650] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 638.158650] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port efa75200-e931-43ca-b12e-4fb50fd45b58, please check neutron logs for more information. [ 638.158650] env[61906]: ERROR nova.compute.manager [ 638.158650] env[61906]: Traceback (most recent call last): [ 638.158650] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 638.158650] env[61906]: listener.cb(fileno) [ 638.158650] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.158650] env[61906]: result = function(*args, **kwargs) [ 638.158650] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.158650] env[61906]: return func(*args, **kwargs) [ 638.158650] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.158650] env[61906]: raise e [ 638.158650] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.158650] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 638.158650] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.158650] env[61906]: created_port_ids = self._update_ports_for_instance( [ 638.158650] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.158650] env[61906]: with excutils.save_and_reraise_exception(): [ 638.158650] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.158650] env[61906]: self.force_reraise() [ 638.158650] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.158650] env[61906]: raise self.value [ 638.158650] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.158650] env[61906]: updated_port = self._update_port( [ 638.158650] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.158650] env[61906]: _ensure_no_port_binding_failure(port) [ 638.158650] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.158650] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 638.159697] env[61906]: nova.exception.PortBindingFailed: Binding failed for port efa75200-e931-43ca-b12e-4fb50fd45b58, please check neutron logs for more information. [ 638.159697] env[61906]: Removing descriptor: 17 [ 638.223633] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.934s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.224276] env[61906]: ERROR nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9aab81df-f193-4ebc-a8ca-e4b8892b9597, please check neutron logs for more information. [ 638.224276] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Traceback (most recent call last): [ 638.224276] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.224276] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] self.driver.spawn(context, instance, image_meta, [ 638.224276] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 638.224276] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.224276] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.224276] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] vm_ref = self.build_virtual_machine(instance, [ 638.224276] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.224276] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.224276] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.224617] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] for vif in network_info: [ 638.224617] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.224617] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] return self._sync_wrapper(fn, *args, **kwargs) [ 638.224617] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.224617] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] self.wait() [ 638.224617] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.224617] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] self[:] = self._gt.wait() [ 638.224617] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.224617] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] return self._exit_event.wait() [ 638.224617] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.224617] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] result = hub.switch() [ 638.224617] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.224617] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] return self.greenlet.switch() [ 638.224963] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.224963] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] result = function(*args, **kwargs) [ 638.224963] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.224963] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] return func(*args, **kwargs) [ 638.224963] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.224963] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] raise e [ 638.224963] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.224963] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] nwinfo = self.network_api.allocate_for_instance( [ 638.224963] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.224963] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] created_port_ids = self._update_ports_for_instance( [ 638.224963] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.224963] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] with excutils.save_and_reraise_exception(): [ 638.224963] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.225342] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] self.force_reraise() [ 638.225342] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.225342] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] raise self.value [ 638.225342] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.225342] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] updated_port = self._update_port( [ 638.225342] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.225342] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] _ensure_no_port_binding_failure(port) [ 638.225342] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.225342] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] raise exception.PortBindingFailed(port_id=port['id']) [ 638.225342] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] nova.exception.PortBindingFailed: Binding failed for port 9aab81df-f193-4ebc-a8ca-e4b8892b9597, please check neutron logs for more information. [ 638.225342] env[61906]: ERROR nova.compute.manager [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] [ 638.225635] env[61906]: DEBUG nova.compute.utils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Binding failed for port 9aab81df-f193-4ebc-a8ca-e4b8892b9597, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 638.226517] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.344s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.229323] env[61906]: DEBUG nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Build of instance b56dd330-d31c-4e42-baf9-d9897c55e633 was re-scheduled: Binding failed for port 9aab81df-f193-4ebc-a8ca-e4b8892b9597, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 638.229323] env[61906]: DEBUG nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 638.229556] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquiring lock "refresh_cache-b56dd330-d31c-4e42-baf9-d9897c55e633" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.229677] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Acquired lock "refresh_cache-b56dd330-d31c-4e42-baf9-d9897c55e633" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.229819] env[61906]: DEBUG nova.network.neutron [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.309999] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 638.336960] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 638.337444] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 638.337734] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 638.341028] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 638.341028] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 638.341028] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 638.341028] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 638.341028] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 638.341385] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 638.341385] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 638.341385] env[61906]: DEBUG nova.virt.hardware [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 638.341385] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940ebbd5-aee5-4bd0-9804-791333850ade {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.350363] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ba0b06-bcc2-49d6-935e-085ffbba3433 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.368120] env[61906]: ERROR nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port efa75200-e931-43ca-b12e-4fb50fd45b58, please check neutron logs for more information. [ 638.368120] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Traceback (most recent call last): [ 638.368120] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 638.368120] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] yield resources [ 638.368120] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.368120] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] self.driver.spawn(context, instance, image_meta, [ 638.368120] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 638.368120] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.368120] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.368120] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] vm_ref = self.build_virtual_machine(instance, [ 638.368120] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.368490] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.368490] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.368490] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] for vif in network_info: [ 638.368490] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.368490] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] return self._sync_wrapper(fn, *args, **kwargs) [ 638.368490] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.368490] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] self.wait() [ 638.368490] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.368490] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] self[:] = self._gt.wait() [ 638.368490] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.368490] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] return self._exit_event.wait() [ 638.368490] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 638.368490] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] current.throw(*self._exc) [ 638.368901] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.368901] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] result = function(*args, **kwargs) [ 638.368901] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.368901] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] return func(*args, **kwargs) [ 638.368901] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.368901] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] raise e [ 638.368901] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.368901] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] nwinfo = self.network_api.allocate_for_instance( [ 638.368901] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.368901] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] created_port_ids = self._update_ports_for_instance( [ 638.368901] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.368901] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] with excutils.save_and_reraise_exception(): [ 638.368901] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.369288] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] self.force_reraise() [ 638.369288] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.369288] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] raise self.value [ 638.369288] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.369288] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] updated_port = self._update_port( [ 638.369288] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.369288] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] _ensure_no_port_binding_failure(port) [ 638.369288] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.369288] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] raise exception.PortBindingFailed(port_id=port['id']) [ 638.369288] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] nova.exception.PortBindingFailed: Binding failed for port efa75200-e931-43ca-b12e-4fb50fd45b58, please check neutron logs for more information. [ 638.369288] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] [ 638.369288] env[61906]: INFO nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Terminating instance [ 638.372517] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "refresh_cache-e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.372826] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquired lock "refresh_cache-e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.373276] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.582604] env[61906]: DEBUG nova.network.neutron [-] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.758400] env[61906]: DEBUG nova.network.neutron [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.890863] env[61906]: DEBUG nova.network.neutron [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.906105] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.013850] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.084905] env[61906]: INFO nova.compute.manager [-] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Took 1.03 seconds to deallocate network for instance. [ 639.089019] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96e9aa2-2730-4918-a4a2-368d37fd4d4b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.091747] env[61906]: DEBUG nova.compute.claims [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 639.091939] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.095897] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d791e3-757a-4e42-be79-12ca764c85fd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.127898] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8241aa-b277-460d-ad0b-d1d484980dd9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.136237] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bef3691-a497-4b6e-a14d-b4d33af89b31 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.150906] env[61906]: DEBUG nova.compute.provider_tree [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.394038] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Releasing lock "refresh_cache-b56dd330-d31c-4e42-baf9-d9897c55e633" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.394038] env[61906]: DEBUG nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 639.394038] env[61906]: DEBUG nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.394292] env[61906]: DEBUG nova.network.neutron [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.411040] env[61906]: DEBUG nova.network.neutron [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.517560] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Releasing lock "refresh_cache-e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.517560] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 639.517560] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 639.517784] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa90020f-0d4b-41b3-bd9e-8e0e310f4123 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.527083] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e235db8b-46b8-47a5-a1f1-2e9b343d9e47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.551692] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e8a2267e-ae4e-4f73-9691-3163ba3ddaa7 could not be found. [ 639.551905] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 639.552097] env[61906]: INFO nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 639.552341] env[61906]: DEBUG oslo.service.loopingcall [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.552558] env[61906]: DEBUG nova.compute.manager [-] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.552647] env[61906]: DEBUG nova.network.neutron [-] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.567345] env[61906]: DEBUG nova.network.neutron [-] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.584938] env[61906]: DEBUG nova.compute.manager [req-8a129f63-8a07-4f7b-80fd-5a0cc81c7903 req-cd35bf78-1a33-4dd1-a9eb-e18de26ea583 service nova] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Received event network-changed-efa75200-e931-43ca-b12e-4fb50fd45b58 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.584938] env[61906]: DEBUG nova.compute.manager [req-8a129f63-8a07-4f7b-80fd-5a0cc81c7903 req-cd35bf78-1a33-4dd1-a9eb-e18de26ea583 service nova] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Refreshing instance network info cache due to event network-changed-efa75200-e931-43ca-b12e-4fb50fd45b58. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 639.585263] env[61906]: DEBUG oslo_concurrency.lockutils [req-8a129f63-8a07-4f7b-80fd-5a0cc81c7903 req-cd35bf78-1a33-4dd1-a9eb-e18de26ea583 service nova] Acquiring lock "refresh_cache-e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.585263] env[61906]: DEBUG oslo_concurrency.lockutils [req-8a129f63-8a07-4f7b-80fd-5a0cc81c7903 req-cd35bf78-1a33-4dd1-a9eb-e18de26ea583 service nova] Acquired lock "refresh_cache-e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.585441] env[61906]: DEBUG nova.network.neutron [req-8a129f63-8a07-4f7b-80fd-5a0cc81c7903 req-cd35bf78-1a33-4dd1-a9eb-e18de26ea583 service nova] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Refreshing network info cache for port efa75200-e931-43ca-b12e-4fb50fd45b58 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 639.654967] env[61906]: DEBUG nova.scheduler.client.report [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.682641] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "ce3c082c-a3cb-452d-9581-0bed6aea892b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.682900] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "ce3c082c-a3cb-452d-9581-0bed6aea892b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.914086] env[61906]: DEBUG nova.network.neutron [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.079559] env[61906]: DEBUG nova.network.neutron [-] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.106777] env[61906]: DEBUG nova.network.neutron [req-8a129f63-8a07-4f7b-80fd-5a0cc81c7903 req-cd35bf78-1a33-4dd1-a9eb-e18de26ea583 service nova] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.161522] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.935s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.162482] env[61906]: ERROR nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0e98277e-b125-4b5e-8318-627224b13767, please check neutron logs for more information. [ 640.162482] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Traceback (most recent call last): [ 640.162482] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.162482] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] self.driver.spawn(context, instance, image_meta, [ 640.162482] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 640.162482] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.162482] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.162482] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] vm_ref = self.build_virtual_machine(instance, [ 640.162482] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.162482] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.162482] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.162845] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] for vif in network_info: [ 640.162845] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.162845] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] return self._sync_wrapper(fn, *args, **kwargs) [ 640.162845] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.162845] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] self.wait() [ 640.162845] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.162845] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] self[:] = self._gt.wait() [ 640.162845] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.162845] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] return self._exit_event.wait() [ 640.162845] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.162845] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] result = hub.switch() [ 640.162845] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.162845] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] return self.greenlet.switch() [ 640.163171] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.163171] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] result = function(*args, **kwargs) [ 640.163171] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 640.163171] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] return func(*args, **kwargs) [ 640.163171] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.163171] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] raise e [ 640.163171] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.163171] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] nwinfo = self.network_api.allocate_for_instance( [ 640.163171] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.163171] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] created_port_ids = self._update_ports_for_instance( [ 640.163171] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.163171] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] with excutils.save_and_reraise_exception(): [ 640.163171] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.163906] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] self.force_reraise() [ 640.163906] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.163906] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] raise self.value [ 640.163906] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.163906] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] updated_port = self._update_port( [ 640.163906] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.163906] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] _ensure_no_port_binding_failure(port) [ 640.163906] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.163906] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] raise exception.PortBindingFailed(port_id=port['id']) [ 640.163906] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] nova.exception.PortBindingFailed: Binding failed for port 0e98277e-b125-4b5e-8318-627224b13767, please check neutron logs for more information. [ 640.163906] env[61906]: ERROR nova.compute.manager [instance: f9edb738-210e-4f85-8062-7759fa9033ea] [ 640.164330] env[61906]: DEBUG nova.compute.utils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Binding failed for port 0e98277e-b125-4b5e-8318-627224b13767, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 640.164376] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.510s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.167610] env[61906]: DEBUG nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Build of instance f9edb738-210e-4f85-8062-7759fa9033ea was re-scheduled: Binding failed for port 0e98277e-b125-4b5e-8318-627224b13767, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 640.168040] env[61906]: DEBUG nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 640.168264] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquiring lock "refresh_cache-f9edb738-210e-4f85-8062-7759fa9033ea" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.168419] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Acquired lock "refresh_cache-f9edb738-210e-4f85-8062-7759fa9033ea" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.168573] env[61906]: DEBUG nova.network.neutron [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.215693] env[61906]: DEBUG nova.network.neutron [req-8a129f63-8a07-4f7b-80fd-5a0cc81c7903 req-cd35bf78-1a33-4dd1-a9eb-e18de26ea583 service nova] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.416763] env[61906]: INFO nova.compute.manager [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] [instance: b56dd330-d31c-4e42-baf9-d9897c55e633] Took 1.02 seconds to deallocate network for instance. [ 640.579798] env[61906]: INFO nova.compute.manager [-] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Took 1.03 seconds to deallocate network for instance. [ 640.580272] env[61906]: DEBUG nova.compute.claims [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 640.580442] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.690736] env[61906]: DEBUG nova.network.neutron [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.718772] env[61906]: DEBUG oslo_concurrency.lockutils [req-8a129f63-8a07-4f7b-80fd-5a0cc81c7903 req-cd35bf78-1a33-4dd1-a9eb-e18de26ea583 service nova] Releasing lock "refresh_cache-e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.718772] env[61906]: DEBUG nova.compute.manager [req-8a129f63-8a07-4f7b-80fd-5a0cc81c7903 req-cd35bf78-1a33-4dd1-a9eb-e18de26ea583 service nova] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Received event network-vif-deleted-efa75200-e931-43ca-b12e-4fb50fd45b58 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 640.736255] env[61906]: DEBUG nova.network.neutron [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.970624] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74221990-2e5d-43e3-b457-192042f7f512 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.978250] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-debaa131-4b44-4ee3-ac49-268896887733 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.010474] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b15ea2-eb6f-4a80-b879-2f80ac41c7be {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.017796] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181cd6c9-cc25-4719-8d97-82cecde9b090 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.031186] env[61906]: DEBUG nova.compute.provider_tree [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.239696] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Releasing lock "refresh_cache-f9edb738-210e-4f85-8062-7759fa9033ea" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.239948] env[61906]: DEBUG nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 641.240141] env[61906]: DEBUG nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.240311] env[61906]: DEBUG nova.network.neutron [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.254763] env[61906]: DEBUG nova.network.neutron [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.444937] env[61906]: INFO nova.scheduler.client.report [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Deleted allocations for instance b56dd330-d31c-4e42-baf9-d9897c55e633 [ 641.534347] env[61906]: DEBUG nova.scheduler.client.report [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.759403] env[61906]: DEBUG nova.network.neutron [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.956198] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9c76b56a-2398-4c94-8c37-af44f7d43472 tempest-ListServerFiltersTestJSON-638303160 tempest-ListServerFiltersTestJSON-638303160-project-member] Lock "b56dd330-d31c-4e42-baf9-d9897c55e633" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.115s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.039419] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.875s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.040080] env[61906]: ERROR nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port af963971-9ac5-40d3-930b-b53aea6dc233, please check neutron logs for more information. [ 642.040080] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Traceback (most recent call last): [ 642.040080] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.040080] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] self.driver.spawn(context, instance, image_meta, [ 642.040080] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 642.040080] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.040080] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.040080] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] vm_ref = self.build_virtual_machine(instance, [ 642.040080] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.040080] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.040080] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.040379] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] for vif in network_info: [ 642.040379] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.040379] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] return self._sync_wrapper(fn, *args, **kwargs) [ 642.040379] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.040379] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] self.wait() [ 642.040379] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.040379] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] self[:] = self._gt.wait() [ 642.040379] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.040379] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] return self._exit_event.wait() [ 642.040379] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 642.040379] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] result = hub.switch() [ 642.040379] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 642.040379] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] return self.greenlet.switch() [ 642.040736] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.040736] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] result = function(*args, **kwargs) [ 642.040736] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.040736] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] return func(*args, **kwargs) [ 642.040736] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.040736] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] raise e [ 642.040736] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.040736] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] nwinfo = self.network_api.allocate_for_instance( [ 642.040736] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.040736] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] created_port_ids = self._update_ports_for_instance( [ 642.040736] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.040736] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] with excutils.save_and_reraise_exception(): [ 642.040736] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.041067] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] self.force_reraise() [ 642.041067] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.041067] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] raise self.value [ 642.041067] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.041067] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] updated_port = self._update_port( [ 642.041067] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.041067] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] _ensure_no_port_binding_failure(port) [ 642.041067] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.041067] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] raise exception.PortBindingFailed(port_id=port['id']) [ 642.041067] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] nova.exception.PortBindingFailed: Binding failed for port af963971-9ac5-40d3-930b-b53aea6dc233, please check neutron logs for more information. [ 642.041067] env[61906]: ERROR nova.compute.manager [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] [ 642.041400] env[61906]: DEBUG nova.compute.utils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Binding failed for port af963971-9ac5-40d3-930b-b53aea6dc233, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 642.042060] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.830s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.043755] env[61906]: DEBUG nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Build of instance 3e6732aa-a878-4bde-9299-abd779f2a109 was re-scheduled: Binding failed for port af963971-9ac5-40d3-930b-b53aea6dc233, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 642.044177] env[61906]: DEBUG nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 642.044539] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Acquiring lock "refresh_cache-3e6732aa-a878-4bde-9299-abd779f2a109" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.044744] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Acquired lock "refresh_cache-3e6732aa-a878-4bde-9299-abd779f2a109" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.044916] env[61906]: DEBUG nova.network.neutron [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 642.261645] env[61906]: INFO nova.compute.manager [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] [instance: f9edb738-210e-4f85-8062-7759fa9033ea] Took 1.02 seconds to deallocate network for instance. [ 642.460029] env[61906]: DEBUG nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 642.564928] env[61906]: DEBUG nova.network.neutron [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.653782] env[61906]: DEBUG nova.network.neutron [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.981528] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.156643] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Releasing lock "refresh_cache-3e6732aa-a878-4bde-9299-abd779f2a109" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.156921] env[61906]: DEBUG nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 643.157105] env[61906]: DEBUG nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.157280] env[61906]: DEBUG nova.network.neutron [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.174682] env[61906]: DEBUG nova.network.neutron [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.296331] env[61906]: INFO nova.scheduler.client.report [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Deleted allocations for instance f9edb738-210e-4f85-8062-7759fa9033ea [ 643.573729] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance f9edb738-210e-4f85-8062-7759fa9033ea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.677763] env[61906]: DEBUG nova.network.neutron [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.805293] env[61906]: DEBUG oslo_concurrency.lockutils [None req-96b1be40-e341-4e1f-9cb7-6a41e9754980 tempest-MigrationsAdminTest-1478707561 tempest-MigrationsAdminTest-1478707561-project-member] Lock "f9edb738-210e-4f85-8062-7759fa9033ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.564s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.076678] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 3e6732aa-a878-4bde-9299-abd779f2a109 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.076923] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 73a96c0e-a08b-46c9-b700-25544e17c73d actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 644.076998] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 0c7d1af9-54e9-435d-a0cc-687f793d4f43 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 644.077135] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 7aee8304-8d07-4aa2-867a-2eeeeacbabc8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 644.077259] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance e8a2267e-ae4e-4f73-9691-3163ba3ddaa7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 644.181012] env[61906]: INFO nova.compute.manager [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] [instance: 3e6732aa-a878-4bde-9299-abd779f2a109] Took 1.02 seconds to deallocate network for instance. [ 644.307669] env[61906]: DEBUG nova.compute.manager [None req-e7371668-f2ba-4576-8b5e-e2856fae326f tempest-ServersListShow296Test-1608988043 tempest-ServersListShow296Test-1608988043-project-member] [instance: 972379ac-0384-4459-8da4-54d747d1f946] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 644.579942] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 569d7ab1-22b8-4006-bf0e-ff35032cb9a6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.812681] env[61906]: DEBUG nova.compute.manager [None req-e7371668-f2ba-4576-8b5e-e2856fae326f tempest-ServersListShow296Test-1608988043 tempest-ServersListShow296Test-1608988043-project-member] [instance: 972379ac-0384-4459-8da4-54d747d1f946] Instance disappeared before build. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 645.083948] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance e5c81f8b-aac3-4cf8-9abb-444731f57206 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.217532] env[61906]: INFO nova.scheduler.client.report [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Deleted allocations for instance 3e6732aa-a878-4bde-9299-abd779f2a109 [ 645.337280] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7371668-f2ba-4576-8b5e-e2856fae326f tempest-ServersListShow296Test-1608988043 tempest-ServersListShow296Test-1608988043-project-member] Lock "972379ac-0384-4459-8da4-54d747d1f946" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.691s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.588798] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 95604a75-0ab3-45d9-a235-9067e615e72a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.729794] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9309b615-33de-4ae2-8a73-6f5210be338d tempest-AttachInterfacesV270Test-847696500 tempest-AttachInterfacesV270Test-847696500-project-member] Lock "3e6732aa-a878-4bde-9299-abd779f2a109" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.121s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.841970] env[61906]: DEBUG nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.096211] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 42135fac-aef7-47d9-9afe-5ac2351bb07c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.234912] env[61906]: DEBUG nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.377935] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.604020] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 71fe23a8-1108-4781-9a35-0156cb9d584d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.758109] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.110807] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 8d4c25cc-47d0-42e0-8fe9-49426175af68 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.619165] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 2f5bd366-317e-4c18-956a-35d9cdbae9e2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.124709] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.628679] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 746ec539-5e13-4a9f-af73-26b175ad1e41 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.133524] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 40dae609-e7a5-4af2-84d9-a53f93d95743 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.476689] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Acquiring lock "a8890726-6c64-4e09-a9a4-0726aa2e8c31" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.476930] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Lock "a8890726-6c64-4e09-a9a4-0726aa2e8c31" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.636609] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 59227a6d-e699-43e3-8f40-e8767ffdc938 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.147253] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.654905] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 3c0602a1-ad7d-4f85-b4e0-d36565e28a38 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.159021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 4a2ed7bd-eb49-49a0-ba81-3c29baea6533 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.492195] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Acquiring lock "ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.492451] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Lock "ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.662101] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 0dca3209-d37b-4536-bbf9-72f0418cef8e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.166065] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance b1ec5d55-96a6-42eb-ba54-c4a3904e22cd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.673066] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 83c86407-16b6-4025-9568-0c9c73d20b10 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.177257] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.682722] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance e0ea5eb3-82cf-44f1-91ee-2c2861aefee4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.186509] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance ce3c082c-a3cb-452d-9581-0bed6aea892b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.186509] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 654.186509] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 654.481695] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd2049c6-3e5d-4d9c-af00-5fd310ea7b18 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.488784] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6efd9dcc-817c-4125-acdd-0d88809fe0e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.520058] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afba9ecd-39a3-40f5-9b06-05dc4d4c57a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.526510] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92f038d-20fc-443d-b502-4515af824fce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.540377] env[61906]: DEBUG nova.compute.provider_tree [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.042883] env[61906]: DEBUG nova.scheduler.client.report [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.549049] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 655.549049] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.507s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.549313] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.054s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.550860] env[61906]: INFO nova.compute.claims [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.832532] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d5db3f-2b85-4e8f-8381-5eef7f2f3859 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.840681] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448d8bc6-1101-4dae-ad29-5cf65bef75ca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.870777] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cdd1031-7d05-4755-aea0-5f38da9f499f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.879024] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5dd1ab-6e4f-4b5f-8d23-7d28661d2825 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.891746] env[61906]: DEBUG nova.compute.provider_tree [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.394890] env[61906]: DEBUG nova.scheduler.client.report [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.900990] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.901567] env[61906]: DEBUG nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 657.904313] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.053s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.906531] env[61906]: INFO nova.compute.claims [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 658.411692] env[61906]: DEBUG nova.compute.utils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 658.418222] env[61906]: DEBUG nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 658.418222] env[61906]: DEBUG nova.network.neutron [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 658.524677] env[61906]: DEBUG nova.policy [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca80129d492e4d68b96ef14a63336ef3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb85e88fd4f54e0db2ff131f81137f64', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 658.853677] env[61906]: DEBUG nova.network.neutron [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Successfully created port: 4980c357-94cf-4788-a94c-3d0d0377dd59 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 658.916358] env[61906]: DEBUG nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 659.305693] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b6e3b8-567f-4d75-a008-d7731a923022 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.314182] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6952bde-8026-430d-926b-e7ac914ecdc4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.346618] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3533597-e7b6-4301-9fe9-a90989913d45 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.355613] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7de6a4f-41bf-4106-8e52-49936e287236 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.369889] env[61906]: DEBUG nova.compute.provider_tree [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.682739] env[61906]: DEBUG nova.compute.manager [req-dea50fa6-6a28-456c-9ad9-658f8d3be8d2 req-34833c47-467f-4b8a-a45a-16246856cd83 service nova] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Received event network-changed-4980c357-94cf-4788-a94c-3d0d0377dd59 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 659.684168] env[61906]: DEBUG nova.compute.manager [req-dea50fa6-6a28-456c-9ad9-658f8d3be8d2 req-34833c47-467f-4b8a-a45a-16246856cd83 service nova] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Refreshing instance network info cache due to event network-changed-4980c357-94cf-4788-a94c-3d0d0377dd59. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 659.684168] env[61906]: DEBUG oslo_concurrency.lockutils [req-dea50fa6-6a28-456c-9ad9-658f8d3be8d2 req-34833c47-467f-4b8a-a45a-16246856cd83 service nova] Acquiring lock "refresh_cache-569d7ab1-22b8-4006-bf0e-ff35032cb9a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.684168] env[61906]: DEBUG oslo_concurrency.lockutils [req-dea50fa6-6a28-456c-9ad9-658f8d3be8d2 req-34833c47-467f-4b8a-a45a-16246856cd83 service nova] Acquired lock "refresh_cache-569d7ab1-22b8-4006-bf0e-ff35032cb9a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.684168] env[61906]: DEBUG nova.network.neutron [req-dea50fa6-6a28-456c-9ad9-658f8d3be8d2 req-34833c47-467f-4b8a-a45a-16246856cd83 service nova] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Refreshing network info cache for port 4980c357-94cf-4788-a94c-3d0d0377dd59 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 659.873602] env[61906]: DEBUG nova.scheduler.client.report [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.910878] env[61906]: ERROR nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4980c357-94cf-4788-a94c-3d0d0377dd59, please check neutron logs for more information. [ 659.910878] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 659.910878] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.910878] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 659.910878] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.910878] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 659.910878] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.910878] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 659.910878] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.910878] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 659.910878] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.910878] env[61906]: ERROR nova.compute.manager raise self.value [ 659.910878] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.910878] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 659.910878] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.910878] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 659.911381] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.911381] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 659.911381] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4980c357-94cf-4788-a94c-3d0d0377dd59, please check neutron logs for more information. [ 659.911381] env[61906]: ERROR nova.compute.manager [ 659.911381] env[61906]: Traceback (most recent call last): [ 659.911381] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 659.911381] env[61906]: listener.cb(fileno) [ 659.911381] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.911381] env[61906]: result = function(*args, **kwargs) [ 659.911381] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 659.911381] env[61906]: return func(*args, **kwargs) [ 659.911381] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.911381] env[61906]: raise e [ 659.911381] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.911381] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 659.911381] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.911381] env[61906]: created_port_ids = self._update_ports_for_instance( [ 659.911381] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.911381] env[61906]: with excutils.save_and_reraise_exception(): [ 659.911381] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.911381] env[61906]: self.force_reraise() [ 659.911381] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.911381] env[61906]: raise self.value [ 659.911381] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.911381] env[61906]: updated_port = self._update_port( [ 659.911381] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.911381] env[61906]: _ensure_no_port_binding_failure(port) [ 659.911381] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.911381] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 659.912333] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 4980c357-94cf-4788-a94c-3d0d0377dd59, please check neutron logs for more information. [ 659.912333] env[61906]: Removing descriptor: 15 [ 659.930211] env[61906]: DEBUG nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 659.957327] env[61906]: DEBUG nova.virt.hardware [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 659.957588] env[61906]: DEBUG nova.virt.hardware [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 659.957747] env[61906]: DEBUG nova.virt.hardware [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 659.957928] env[61906]: DEBUG nova.virt.hardware [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 659.958088] env[61906]: DEBUG nova.virt.hardware [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 659.958241] env[61906]: DEBUG nova.virt.hardware [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 659.958453] env[61906]: DEBUG nova.virt.hardware [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 659.958613] env[61906]: DEBUG nova.virt.hardware [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 659.958779] env[61906]: DEBUG nova.virt.hardware [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 659.958943] env[61906]: DEBUG nova.virt.hardware [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 659.959145] env[61906]: DEBUG nova.virt.hardware [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 659.960058] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a56c02e-b07e-4392-b0cd-cb159706bc9a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.968605] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52fe2d23-9427-4058-9dfa-8870cdbf479c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.982608] env[61906]: ERROR nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4980c357-94cf-4788-a94c-3d0d0377dd59, please check neutron logs for more information. [ 659.982608] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Traceback (most recent call last): [ 659.982608] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 659.982608] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] yield resources [ 659.982608] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 659.982608] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] self.driver.spawn(context, instance, image_meta, [ 659.982608] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 659.982608] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.982608] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.982608] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] vm_ref = self.build_virtual_machine(instance, [ 659.982608] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.982957] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.982957] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.982957] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] for vif in network_info: [ 659.982957] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.982957] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] return self._sync_wrapper(fn, *args, **kwargs) [ 659.982957] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.982957] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] self.wait() [ 659.982957] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.982957] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] self[:] = self._gt.wait() [ 659.982957] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.982957] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] return self._exit_event.wait() [ 659.982957] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 659.982957] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] current.throw(*self._exc) [ 659.983614] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.983614] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] result = function(*args, **kwargs) [ 659.983614] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 659.983614] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] return func(*args, **kwargs) [ 659.983614] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.983614] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] raise e [ 659.983614] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.983614] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] nwinfo = self.network_api.allocate_for_instance( [ 659.983614] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.983614] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] created_port_ids = self._update_ports_for_instance( [ 659.983614] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.983614] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] with excutils.save_and_reraise_exception(): [ 659.983614] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.984411] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] self.force_reraise() [ 659.984411] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.984411] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] raise self.value [ 659.984411] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.984411] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] updated_port = self._update_port( [ 659.984411] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.984411] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] _ensure_no_port_binding_failure(port) [ 659.984411] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.984411] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] raise exception.PortBindingFailed(port_id=port['id']) [ 659.984411] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] nova.exception.PortBindingFailed: Binding failed for port 4980c357-94cf-4788-a94c-3d0d0377dd59, please check neutron logs for more information. [ 659.984411] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] [ 659.984411] env[61906]: INFO nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Terminating instance [ 659.985317] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "refresh_cache-569d7ab1-22b8-4006-bf0e-ff35032cb9a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.201483] env[61906]: DEBUG nova.network.neutron [req-dea50fa6-6a28-456c-9ad9-658f8d3be8d2 req-34833c47-467f-4b8a-a45a-16246856cd83 service nova] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.280227] env[61906]: DEBUG nova.network.neutron [req-dea50fa6-6a28-456c-9ad9-658f8d3be8d2 req-34833c47-467f-4b8a-a45a-16246856cd83 service nova] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.381041] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.381189] env[61906]: DEBUG nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 660.385629] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.523s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.783057] env[61906]: DEBUG oslo_concurrency.lockutils [req-dea50fa6-6a28-456c-9ad9-658f8d3be8d2 req-34833c47-467f-4b8a-a45a-16246856cd83 service nova] Releasing lock "refresh_cache-569d7ab1-22b8-4006-bf0e-ff35032cb9a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.783376] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "refresh_cache-569d7ab1-22b8-4006-bf0e-ff35032cb9a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.783523] env[61906]: DEBUG nova.network.neutron [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 660.890406] env[61906]: DEBUG nova.compute.utils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 660.894943] env[61906]: DEBUG nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 660.895136] env[61906]: DEBUG nova.network.neutron [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 660.949426] env[61906]: DEBUG nova.policy [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '70a294a034cf41fa93084149243c880e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '256a7dcc6f1549b58e81c41509c0500e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 661.235194] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-491c1dc8-d685-4808-8db5-34d1d55dea00 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.244174] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160f27d6-4ee7-4828-891d-a00510193bd0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.275145] env[61906]: DEBUG nova.network.neutron [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Successfully created port: 16f2305d-3260-4121-9d5c-89b1cf66ccf0 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 661.277668] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55287888-c252-48ce-80ee-79fab3bb3488 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.285897] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4388495e-1908-41d1-91df-6f5cb31c474f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.302774] env[61906]: DEBUG nova.compute.provider_tree [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.307396] env[61906]: DEBUG nova.network.neutron [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.388562] env[61906]: DEBUG nova.network.neutron [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.397422] env[61906]: DEBUG nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 661.734658] env[61906]: DEBUG nova.compute.manager [req-efeb85b2-1cbd-4e41-850e-116ea7dab3cb req-e9a5fe40-79fd-47ef-8e63-bcca8d964402 service nova] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Received event network-vif-deleted-4980c357-94cf-4788-a94c-3d0d0377dd59 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 661.806130] env[61906]: DEBUG nova.scheduler.client.report [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.890314] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "refresh_cache-569d7ab1-22b8-4006-bf0e-ff35032cb9a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.890720] env[61906]: DEBUG nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 661.890911] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 661.891658] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-85a47715-df0e-40cd-97a2-3786221018a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.901798] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a479b169-19c6-4c48-b6cd-c928c1d779ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.929119] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 569d7ab1-22b8-4006-bf0e-ff35032cb9a6 could not be found. [ 661.929399] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 661.929682] env[61906]: INFO nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 661.930812] env[61906]: DEBUG oslo.service.loopingcall [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 661.931120] env[61906]: DEBUG nova.compute.manager [-] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.931242] env[61906]: DEBUG nova.network.neutron [-] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 661.947939] env[61906]: DEBUG nova.network.neutron [-] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.315165] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.926s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.315165] env[61906]: ERROR nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port def891b7-a3b4-4b8e-bdbe-9e50e2f817c3, please check neutron logs for more information. [ 662.315165] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Traceback (most recent call last): [ 662.315165] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 662.315165] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] self.driver.spawn(context, instance, image_meta, [ 662.315165] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 662.315165] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.315165] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.315165] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] vm_ref = self.build_virtual_machine(instance, [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] for vif in network_info: [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] return self._sync_wrapper(fn, *args, **kwargs) [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] self.wait() [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] self[:] = self._gt.wait() [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] return self._exit_event.wait() [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 662.315581] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] current.throw(*self._exc) [ 662.315971] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.315971] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] result = function(*args, **kwargs) [ 662.315971] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 662.315971] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] return func(*args, **kwargs) [ 662.315971] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.315971] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] raise e [ 662.315971] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.315971] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] nwinfo = self.network_api.allocate_for_instance( [ 662.315971] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.315971] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] created_port_ids = self._update_ports_for_instance( [ 662.315971] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.315971] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] with excutils.save_and_reraise_exception(): [ 662.315971] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.316340] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] self.force_reraise() [ 662.316340] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.316340] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] raise self.value [ 662.316340] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.316340] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] updated_port = self._update_port( [ 662.316340] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.316340] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] _ensure_no_port_binding_failure(port) [ 662.316340] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.316340] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] raise exception.PortBindingFailed(port_id=port['id']) [ 662.316340] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] nova.exception.PortBindingFailed: Binding failed for port def891b7-a3b4-4b8e-bdbe-9e50e2f817c3, please check neutron logs for more information. [ 662.316340] env[61906]: ERROR nova.compute.manager [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] [ 662.316614] env[61906]: DEBUG nova.compute.utils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Binding failed for port def891b7-a3b4-4b8e-bdbe-9e50e2f817c3, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 662.316614] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.853s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.317642] env[61906]: DEBUG nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Build of instance 73a96c0e-a08b-46c9-b700-25544e17c73d was re-scheduled: Binding failed for port def891b7-a3b4-4b8e-bdbe-9e50e2f817c3, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 662.318131] env[61906]: DEBUG nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 662.318380] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Acquiring lock "refresh_cache-73a96c0e-a08b-46c9-b700-25544e17c73d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.318551] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Acquired lock "refresh_cache-73a96c0e-a08b-46c9-b700-25544e17c73d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.318731] env[61906]: DEBUG nova.network.neutron [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 662.320562] env[61906]: ERROR nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 16f2305d-3260-4121-9d5c-89b1cf66ccf0, please check neutron logs for more information. [ 662.320562] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 662.320562] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.320562] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 662.320562] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.320562] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 662.320562] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.320562] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 662.320562] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.320562] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 662.320562] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.320562] env[61906]: ERROR nova.compute.manager raise self.value [ 662.320562] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.320562] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 662.320562] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.320562] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 662.321040] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.321040] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 662.321040] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 16f2305d-3260-4121-9d5c-89b1cf66ccf0, please check neutron logs for more information. [ 662.321040] env[61906]: ERROR nova.compute.manager [ 662.321040] env[61906]: Traceback (most recent call last): [ 662.321040] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 662.321040] env[61906]: listener.cb(fileno) [ 662.321040] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.321040] env[61906]: result = function(*args, **kwargs) [ 662.321040] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 662.321040] env[61906]: return func(*args, **kwargs) [ 662.321040] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.321040] env[61906]: raise e [ 662.321040] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.321040] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 662.321040] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.321040] env[61906]: created_port_ids = self._update_ports_for_instance( [ 662.321040] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.321040] env[61906]: with excutils.save_and_reraise_exception(): [ 662.321040] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.321040] env[61906]: self.force_reraise() [ 662.321040] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.321040] env[61906]: raise self.value [ 662.321040] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.321040] env[61906]: updated_port = self._update_port( [ 662.321040] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.321040] env[61906]: _ensure_no_port_binding_failure(port) [ 662.321040] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.321040] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 662.321805] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 16f2305d-3260-4121-9d5c-89b1cf66ccf0, please check neutron logs for more information. [ 662.321805] env[61906]: Removing descriptor: 15 [ 662.418631] env[61906]: DEBUG nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 662.442557] env[61906]: DEBUG nova.virt.hardware [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 662.442828] env[61906]: DEBUG nova.virt.hardware [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 662.442982] env[61906]: DEBUG nova.virt.hardware [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 662.443175] env[61906]: DEBUG nova.virt.hardware [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 662.443349] env[61906]: DEBUG nova.virt.hardware [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 662.443506] env[61906]: DEBUG nova.virt.hardware [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 662.443715] env[61906]: DEBUG nova.virt.hardware [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 662.443872] env[61906]: DEBUG nova.virt.hardware [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 662.444136] env[61906]: DEBUG nova.virt.hardware [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 662.444368] env[61906]: DEBUG nova.virt.hardware [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 662.444590] env[61906]: DEBUG nova.virt.hardware [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 662.446366] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3859dc1f-282c-4737-acfb-4c0b7cff9958 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.450599] env[61906]: DEBUG nova.network.neutron [-] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.455479] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba2e6d3-096d-4a72-a96f-eb2d14faabff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.469799] env[61906]: ERROR nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 16f2305d-3260-4121-9d5c-89b1cf66ccf0, please check neutron logs for more information. [ 662.469799] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Traceback (most recent call last): [ 662.469799] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 662.469799] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] yield resources [ 662.469799] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 662.469799] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] self.driver.spawn(context, instance, image_meta, [ 662.469799] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 662.469799] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.469799] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.469799] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] vm_ref = self.build_virtual_machine(instance, [ 662.469799] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.470241] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.470241] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.470241] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] for vif in network_info: [ 662.470241] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.470241] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] return self._sync_wrapper(fn, *args, **kwargs) [ 662.470241] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.470241] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] self.wait() [ 662.470241] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.470241] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] self[:] = self._gt.wait() [ 662.470241] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.470241] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] return self._exit_event.wait() [ 662.470241] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 662.470241] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] current.throw(*self._exc) [ 662.470796] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.470796] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] result = function(*args, **kwargs) [ 662.470796] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 662.470796] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] return func(*args, **kwargs) [ 662.470796] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.470796] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] raise e [ 662.470796] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.470796] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] nwinfo = self.network_api.allocate_for_instance( [ 662.470796] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.470796] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] created_port_ids = self._update_ports_for_instance( [ 662.470796] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.470796] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] with excutils.save_and_reraise_exception(): [ 662.470796] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.471535] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] self.force_reraise() [ 662.471535] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.471535] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] raise self.value [ 662.471535] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.471535] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] updated_port = self._update_port( [ 662.471535] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.471535] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] _ensure_no_port_binding_failure(port) [ 662.471535] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.471535] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] raise exception.PortBindingFailed(port_id=port['id']) [ 662.471535] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] nova.exception.PortBindingFailed: Binding failed for port 16f2305d-3260-4121-9d5c-89b1cf66ccf0, please check neutron logs for more information. [ 662.471535] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] [ 662.471535] env[61906]: INFO nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Terminating instance [ 662.472915] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquiring lock "refresh_cache-e5c81f8b-aac3-4cf8-9abb-444731f57206" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.473086] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquired lock "refresh_cache-e5c81f8b-aac3-4cf8-9abb-444731f57206" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.473254] env[61906]: DEBUG nova.network.neutron [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 662.839556] env[61906]: DEBUG nova.network.neutron [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.937515] env[61906]: DEBUG nova.network.neutron [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.954825] env[61906]: INFO nova.compute.manager [-] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Took 1.02 seconds to deallocate network for instance. [ 662.959074] env[61906]: DEBUG nova.compute.claims [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 662.959268] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.992142] env[61906]: DEBUG nova.network.neutron [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.076146] env[61906]: DEBUG nova.network.neutron [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.139673] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493941b1-2479-4b32-b014-ccfd7fa13508 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.150086] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82138f1a-b4f7-4d2f-83bf-9aa20fb0b261 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.180885] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-707c1c9c-8e0e-4e50-9703-90cb189e3786 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.189086] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffde9fb7-12de-4a71-ba0d-746ac742080c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.206398] env[61906]: DEBUG nova.compute.provider_tree [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.440723] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Releasing lock "refresh_cache-73a96c0e-a08b-46c9-b700-25544e17c73d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.440998] env[61906]: DEBUG nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 663.441209] env[61906]: DEBUG nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.441408] env[61906]: DEBUG nova.network.neutron [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 663.456367] env[61906]: DEBUG nova.network.neutron [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.579495] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Releasing lock "refresh_cache-e5c81f8b-aac3-4cf8-9abb-444731f57206" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.579983] env[61906]: DEBUG nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 663.580184] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 663.580492] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-297e3c36-da3d-4715-96ac-4862a4c7c35a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.589769] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e525fc-ff8d-4edc-a2ba-255fa3b6df9a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.613305] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e5c81f8b-aac3-4cf8-9abb-444731f57206 could not be found. [ 663.613409] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 663.613593] env[61906]: INFO nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Took 0.03 seconds to destroy the instance on the hypervisor. [ 663.613849] env[61906]: DEBUG oslo.service.loopingcall [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 663.614067] env[61906]: DEBUG nova.compute.manager [-] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.614162] env[61906]: DEBUG nova.network.neutron [-] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 663.634667] env[61906]: DEBUG nova.network.neutron [-] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.709241] env[61906]: DEBUG nova.scheduler.client.report [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.790935] env[61906]: DEBUG nova.compute.manager [req-1e476a0b-213d-4ec8-8495-ac7486aadd3d req-1175a6a8-78fc-42f4-967f-3291424ade68 service nova] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Received event network-changed-16f2305d-3260-4121-9d5c-89b1cf66ccf0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 663.791223] env[61906]: DEBUG nova.compute.manager [req-1e476a0b-213d-4ec8-8495-ac7486aadd3d req-1175a6a8-78fc-42f4-967f-3291424ade68 service nova] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Refreshing instance network info cache due to event network-changed-16f2305d-3260-4121-9d5c-89b1cf66ccf0. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 663.791343] env[61906]: DEBUG oslo_concurrency.lockutils [req-1e476a0b-213d-4ec8-8495-ac7486aadd3d req-1175a6a8-78fc-42f4-967f-3291424ade68 service nova] Acquiring lock "refresh_cache-e5c81f8b-aac3-4cf8-9abb-444731f57206" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.791510] env[61906]: DEBUG oslo_concurrency.lockutils [req-1e476a0b-213d-4ec8-8495-ac7486aadd3d req-1175a6a8-78fc-42f4-967f-3291424ade68 service nova] Acquired lock "refresh_cache-e5c81f8b-aac3-4cf8-9abb-444731f57206" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.791703] env[61906]: DEBUG nova.network.neutron [req-1e476a0b-213d-4ec8-8495-ac7486aadd3d req-1175a6a8-78fc-42f4-967f-3291424ade68 service nova] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Refreshing network info cache for port 16f2305d-3260-4121-9d5c-89b1cf66ccf0 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 663.959068] env[61906]: DEBUG nova.network.neutron [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.137251] env[61906]: DEBUG nova.network.neutron [-] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.216086] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.900s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.216086] env[61906]: ERROR nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 116f6b3d-5d45-4f71-ac3d-df6f8361e722, please check neutron logs for more information. [ 664.216086] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Traceback (most recent call last): [ 664.216086] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.216086] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] self.driver.spawn(context, instance, image_meta, [ 664.216086] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 664.216086] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.216086] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.216086] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] vm_ref = self.build_virtual_machine(instance, [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] for vif in network_info: [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] return self._sync_wrapper(fn, *args, **kwargs) [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] self.wait() [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] self[:] = self._gt.wait() [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] return self._exit_event.wait() [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 664.216723] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] current.throw(*self._exc) [ 664.217520] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.217520] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] result = function(*args, **kwargs) [ 664.217520] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 664.217520] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] return func(*args, **kwargs) [ 664.217520] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.217520] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] raise e [ 664.217520] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.217520] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] nwinfo = self.network_api.allocate_for_instance( [ 664.217520] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.217520] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] created_port_ids = self._update_ports_for_instance( [ 664.217520] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.217520] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] with excutils.save_and_reraise_exception(): [ 664.217520] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.217849] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] self.force_reraise() [ 664.217849] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.217849] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] raise self.value [ 664.217849] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.217849] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] updated_port = self._update_port( [ 664.217849] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.217849] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] _ensure_no_port_binding_failure(port) [ 664.217849] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.217849] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] raise exception.PortBindingFailed(port_id=port['id']) [ 664.217849] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] nova.exception.PortBindingFailed: Binding failed for port 116f6b3d-5d45-4f71-ac3d-df6f8361e722, please check neutron logs for more information. [ 664.217849] env[61906]: ERROR nova.compute.manager [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] [ 664.218217] env[61906]: DEBUG nova.compute.utils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Binding failed for port 116f6b3d-5d45-4f71-ac3d-df6f8361e722, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 664.218217] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.059s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.218663] env[61906]: INFO nova.compute.claims [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.222362] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Build of instance 0c7d1af9-54e9-435d-a0cc-687f793d4f43 was re-scheduled: Binding failed for port 116f6b3d-5d45-4f71-ac3d-df6f8361e722, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 664.222865] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 664.223146] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "refresh_cache-0c7d1af9-54e9-435d-a0cc-687f793d4f43" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.223355] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquired lock "refresh_cache-0c7d1af9-54e9-435d-a0cc-687f793d4f43" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.223533] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 664.322642] env[61906]: DEBUG nova.network.neutron [req-1e476a0b-213d-4ec8-8495-ac7486aadd3d req-1175a6a8-78fc-42f4-967f-3291424ade68 service nova] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.438881] env[61906]: DEBUG nova.network.neutron [req-1e476a0b-213d-4ec8-8495-ac7486aadd3d req-1175a6a8-78fc-42f4-967f-3291424ade68 service nova] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.461816] env[61906]: INFO nova.compute.manager [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] [instance: 73a96c0e-a08b-46c9-b700-25544e17c73d] Took 1.02 seconds to deallocate network for instance. [ 664.640225] env[61906]: INFO nova.compute.manager [-] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Took 1.03 seconds to deallocate network for instance. [ 664.642556] env[61906]: DEBUG nova.compute.claims [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 664.642741] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.747687] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.814232] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.941789] env[61906]: DEBUG oslo_concurrency.lockutils [req-1e476a0b-213d-4ec8-8495-ac7486aadd3d req-1175a6a8-78fc-42f4-967f-3291424ade68 service nova] Releasing lock "refresh_cache-e5c81f8b-aac3-4cf8-9abb-444731f57206" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.942154] env[61906]: DEBUG nova.compute.manager [req-1e476a0b-213d-4ec8-8495-ac7486aadd3d req-1175a6a8-78fc-42f4-967f-3291424ade68 service nova] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Received event network-vif-deleted-16f2305d-3260-4121-9d5c-89b1cf66ccf0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 665.316653] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Releasing lock "refresh_cache-0c7d1af9-54e9-435d-a0cc-687f793d4f43" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.316892] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 665.317091] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 665.317262] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 665.334748] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.487071] env[61906]: INFO nova.scheduler.client.report [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Deleted allocations for instance 73a96c0e-a08b-46c9-b700-25544e17c73d [ 665.504472] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5b0a19-6912-4b58-a0a9-e6ec7226c19d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.513276] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f264d40d-99a3-482f-8843-ee6fcc59a76f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.545882] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3807e738-e756-49b6-a510-4c97cae942a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.554611] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b4f2d6-6070-414c-bb78-1c7968ad09d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.568435] env[61906]: DEBUG nova.compute.provider_tree [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.837603] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.997632] env[61906]: DEBUG oslo_concurrency.lockutils [None req-225b9cce-432c-4ccf-8ecf-a31a758cb025 tempest-ServersTestBootFromVolume-112493301 tempest-ServersTestBootFromVolume-112493301-project-member] Lock "73a96c0e-a08b-46c9-b700-25544e17c73d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.618s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.072054] env[61906]: DEBUG nova.scheduler.client.report [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.341144] env[61906]: INFO nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 0c7d1af9-54e9-435d-a0cc-687f793d4f43] Took 1.02 seconds to deallocate network for instance. [ 666.500031] env[61906]: DEBUG nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 666.575935] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.576515] env[61906]: DEBUG nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 666.579271] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.487s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.037224] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.084177] env[61906]: DEBUG nova.compute.utils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 667.088897] env[61906]: DEBUG nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 667.088897] env[61906]: DEBUG nova.network.neutron [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 667.139406] env[61906]: DEBUG nova.policy [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd4d2def35d334a67b25b36124197a1d2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ec491f4c3ae40d8a60fe37d0dc7b2ef', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 667.380067] env[61906]: INFO nova.scheduler.client.report [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Deleted allocations for instance 0c7d1af9-54e9-435d-a0cc-687f793d4f43 [ 667.454021] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90beb238-0121-4a06-a743-b83e39690e2b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.462358] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ed93a9-ad6f-4a7c-afb4-079fd47316ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.495605] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3deecdd0-7b41-4a3f-be1f-1aad763720bb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.506790] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-541b773a-8aa4-4b0e-8eae-587827327344 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.522632] env[61906]: DEBUG nova.compute.provider_tree [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.537236] env[61906]: DEBUG nova.network.neutron [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Successfully created port: f1de37fc-eff3-425a-a263-b13b723e3323 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 667.595016] env[61906]: DEBUG nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 667.888873] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "0c7d1af9-54e9-435d-a0cc-687f793d4f43" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.318s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.928795] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Acquiring lock "45621a25-cd9c-4931-899c-647da0b50ec7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.928795] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Lock "45621a25-cd9c-4931-899c-647da0b50ec7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.025070] env[61906]: DEBUG nova.scheduler.client.report [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.391751] env[61906]: DEBUG nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 668.530256] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.530927] env[61906]: ERROR nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7df09b9b-9912-4b71-9cd2-6e0183ee2bde, please check neutron logs for more information. [ 668.530927] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Traceback (most recent call last): [ 668.530927] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.530927] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] self.driver.spawn(context, instance, image_meta, [ 668.530927] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 668.530927] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.530927] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.530927] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] vm_ref = self.build_virtual_machine(instance, [ 668.530927] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.530927] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.530927] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] for vif in network_info: [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] return self._sync_wrapper(fn, *args, **kwargs) [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] self.wait() [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] self[:] = self._gt.wait() [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] return self._exit_event.wait() [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] current.throw(*self._exc) [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] result = function(*args, **kwargs) [ 668.531346] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 668.531785] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] return func(*args, **kwargs) [ 668.531785] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.531785] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] raise e [ 668.531785] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.531785] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] nwinfo = self.network_api.allocate_for_instance( [ 668.531785] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.531785] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] created_port_ids = self._update_ports_for_instance( [ 668.531785] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.531785] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] with excutils.save_and_reraise_exception(): [ 668.531785] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.531785] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] self.force_reraise() [ 668.531785] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.531785] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] raise self.value [ 668.532191] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.532191] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] updated_port = self._update_port( [ 668.532191] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.532191] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] _ensure_no_port_binding_failure(port) [ 668.532191] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.532191] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] raise exception.PortBindingFailed(port_id=port['id']) [ 668.532191] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] nova.exception.PortBindingFailed: Binding failed for port 7df09b9b-9912-4b71-9cd2-6e0183ee2bde, please check neutron logs for more information. [ 668.532191] env[61906]: ERROR nova.compute.manager [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] [ 668.532191] env[61906]: DEBUG nova.compute.utils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Binding failed for port 7df09b9b-9912-4b71-9cd2-6e0183ee2bde, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 668.532945] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.952s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.539398] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Build of instance 7aee8304-8d07-4aa2-867a-2eeeeacbabc8 was re-scheduled: Binding failed for port 7df09b9b-9912-4b71-9cd2-6e0183ee2bde, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 668.539831] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 668.540077] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "refresh_cache-7aee8304-8d07-4aa2-867a-2eeeeacbabc8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.540229] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquired lock "refresh_cache-7aee8304-8d07-4aa2-867a-2eeeeacbabc8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.540387] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 668.608270] env[61906]: DEBUG nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 668.628994] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.644451] env[61906]: DEBUG nova.virt.hardware [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.644704] env[61906]: DEBUG nova.virt.hardware [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.644858] env[61906]: DEBUG nova.virt.hardware [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.645233] env[61906]: DEBUG nova.virt.hardware [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.645456] env[61906]: DEBUG nova.virt.hardware [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.645685] env[61906]: DEBUG nova.virt.hardware [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.645963] env[61906]: DEBUG nova.virt.hardware [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.646217] env[61906]: DEBUG nova.virt.hardware [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.646452] env[61906]: DEBUG nova.virt.hardware [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.646673] env[61906]: DEBUG nova.virt.hardware [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.646890] env[61906]: DEBUG nova.virt.hardware [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.647898] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6849d4ed-bf19-4e76-8d19-4f20801f0c10 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.658389] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b43d47-3d11-42c2-bac3-d672ee7f7987 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.924602] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.950044] env[61906]: DEBUG nova.compute.manager [req-e37d7e07-c37e-423f-a5c3-5f00418688d7 req-804d1481-b605-4a4e-8c3d-9e70f03a716e service nova] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Received event network-changed-f1de37fc-eff3-425a-a263-b13b723e3323 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 668.950181] env[61906]: DEBUG nova.compute.manager [req-e37d7e07-c37e-423f-a5c3-5f00418688d7 req-804d1481-b605-4a4e-8c3d-9e70f03a716e service nova] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Refreshing instance network info cache due to event network-changed-f1de37fc-eff3-425a-a263-b13b723e3323. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 668.952203] env[61906]: DEBUG oslo_concurrency.lockutils [req-e37d7e07-c37e-423f-a5c3-5f00418688d7 req-804d1481-b605-4a4e-8c3d-9e70f03a716e service nova] Acquiring lock "refresh_cache-95604a75-0ab3-45d9-a235-9067e615e72a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.952203] env[61906]: DEBUG oslo_concurrency.lockutils [req-e37d7e07-c37e-423f-a5c3-5f00418688d7 req-804d1481-b605-4a4e-8c3d-9e70f03a716e service nova] Acquired lock "refresh_cache-95604a75-0ab3-45d9-a235-9067e615e72a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.952203] env[61906]: DEBUG nova.network.neutron [req-e37d7e07-c37e-423f-a5c3-5f00418688d7 req-804d1481-b605-4a4e-8c3d-9e70f03a716e service nova] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Refreshing network info cache for port f1de37fc-eff3-425a-a263-b13b723e3323 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 669.045264] env[61906]: DEBUG nova.compute.utils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Can not refresh info_cache because instance was not found {{(pid=61906) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 669.047996] env[61906]: ERROR nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f1de37fc-eff3-425a-a263-b13b723e3323, please check neutron logs for more information. [ 669.047996] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 669.047996] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.047996] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 669.047996] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.047996] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 669.047996] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.047996] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 669.047996] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.047996] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 669.047996] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.047996] env[61906]: ERROR nova.compute.manager raise self.value [ 669.047996] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.047996] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 669.047996] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.047996] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 669.048949] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.048949] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 669.048949] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f1de37fc-eff3-425a-a263-b13b723e3323, please check neutron logs for more information. [ 669.048949] env[61906]: ERROR nova.compute.manager [ 669.048949] env[61906]: Traceback (most recent call last): [ 669.048949] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 669.048949] env[61906]: listener.cb(fileno) [ 669.048949] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.048949] env[61906]: result = function(*args, **kwargs) [ 669.048949] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 669.048949] env[61906]: return func(*args, **kwargs) [ 669.048949] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.048949] env[61906]: raise e [ 669.048949] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.048949] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 669.048949] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.048949] env[61906]: created_port_ids = self._update_ports_for_instance( [ 669.048949] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.048949] env[61906]: with excutils.save_and_reraise_exception(): [ 669.048949] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.048949] env[61906]: self.force_reraise() [ 669.048949] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.048949] env[61906]: raise self.value [ 669.048949] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.048949] env[61906]: updated_port = self._update_port( [ 669.048949] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.048949] env[61906]: _ensure_no_port_binding_failure(port) [ 669.048949] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.048949] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 669.050270] env[61906]: nova.exception.PortBindingFailed: Binding failed for port f1de37fc-eff3-425a-a263-b13b723e3323, please check neutron logs for more information. [ 669.050270] env[61906]: Removing descriptor: 15 [ 669.050490] env[61906]: ERROR nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f1de37fc-eff3-425a-a263-b13b723e3323, please check neutron logs for more information. [ 669.050490] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Traceback (most recent call last): [ 669.050490] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 669.050490] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] yield resources [ 669.050490] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.050490] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] self.driver.spawn(context, instance, image_meta, [ 669.050490] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 669.050490] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.050490] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.050490] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] vm_ref = self.build_virtual_machine(instance, [ 669.050490] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.050849] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.050849] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.050849] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] for vif in network_info: [ 669.050849] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.050849] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] return self._sync_wrapper(fn, *args, **kwargs) [ 669.050849] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.050849] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] self.wait() [ 669.050849] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.050849] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] self[:] = self._gt.wait() [ 669.050849] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.050849] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] return self._exit_event.wait() [ 669.050849] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.050849] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] result = hub.switch() [ 669.051683] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.051683] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] return self.greenlet.switch() [ 669.051683] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.051683] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] result = function(*args, **kwargs) [ 669.051683] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 669.051683] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] return func(*args, **kwargs) [ 669.051683] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.051683] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] raise e [ 669.051683] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.051683] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] nwinfo = self.network_api.allocate_for_instance( [ 669.051683] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.051683] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] created_port_ids = self._update_ports_for_instance( [ 669.051683] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.052039] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] with excutils.save_and_reraise_exception(): [ 669.052039] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.052039] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] self.force_reraise() [ 669.052039] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.052039] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] raise self.value [ 669.052039] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.052039] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] updated_port = self._update_port( [ 669.052039] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.052039] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] _ensure_no_port_binding_failure(port) [ 669.052039] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.052039] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] raise exception.PortBindingFailed(port_id=port['id']) [ 669.052039] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] nova.exception.PortBindingFailed: Binding failed for port f1de37fc-eff3-425a-a263-b13b723e3323, please check neutron logs for more information. [ 669.052039] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] [ 669.052371] env[61906]: INFO nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Terminating instance [ 669.054616] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Acquiring lock "refresh_cache-95604a75-0ab3-45d9-a235-9067e615e72a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.068071] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.170833] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.360464] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb64877-3529-42ea-b73c-724d5c2dbc5f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.370816] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ef317a-1b92-4687-a927-947edeed3792 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.400828] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-679b969c-d8bf-49cb-9276-8c238c28b4f3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.408679] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb58cfe5-e220-468b-abce-9e7db83bb1e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.422430] env[61906]: DEBUG nova.compute.provider_tree [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.476903] env[61906]: DEBUG nova.network.neutron [req-e37d7e07-c37e-423f-a5c3-5f00418688d7 req-804d1481-b605-4a4e-8c3d-9e70f03a716e service nova] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.558795] env[61906]: DEBUG nova.network.neutron [req-e37d7e07-c37e-423f-a5c3-5f00418688d7 req-804d1481-b605-4a4e-8c3d-9e70f03a716e service nova] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.674194] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Releasing lock "refresh_cache-7aee8304-8d07-4aa2-867a-2eeeeacbabc8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.674586] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 669.674737] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.674901] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 669.691405] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.930378] env[61906]: DEBUG nova.scheduler.client.report [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.061640] env[61906]: DEBUG oslo_concurrency.lockutils [req-e37d7e07-c37e-423f-a5c3-5f00418688d7 req-804d1481-b605-4a4e-8c3d-9e70f03a716e service nova] Releasing lock "refresh_cache-95604a75-0ab3-45d9-a235-9067e615e72a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.062060] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Acquired lock "refresh_cache-95604a75-0ab3-45d9-a235-9067e615e72a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.062249] env[61906]: DEBUG nova.network.neutron [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 670.194618] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.435955] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.903s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.436655] env[61906]: ERROR nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port efa75200-e931-43ca-b12e-4fb50fd45b58, please check neutron logs for more information. [ 670.436655] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Traceback (most recent call last): [ 670.436655] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.436655] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] self.driver.spawn(context, instance, image_meta, [ 670.436655] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 670.436655] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.436655] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.436655] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] vm_ref = self.build_virtual_machine(instance, [ 670.436655] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.436655] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.436655] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] for vif in network_info: [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] return self._sync_wrapper(fn, *args, **kwargs) [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] self.wait() [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] self[:] = self._gt.wait() [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] return self._exit_event.wait() [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] current.throw(*self._exc) [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] result = function(*args, **kwargs) [ 670.436968] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 670.437360] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] return func(*args, **kwargs) [ 670.437360] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.437360] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] raise e [ 670.437360] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.437360] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] nwinfo = self.network_api.allocate_for_instance( [ 670.437360] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.437360] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] created_port_ids = self._update_ports_for_instance( [ 670.437360] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.437360] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] with excutils.save_and_reraise_exception(): [ 670.437360] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.437360] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] self.force_reraise() [ 670.437360] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.437360] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] raise self.value [ 670.437789] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.437789] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] updated_port = self._update_port( [ 670.437789] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.437789] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] _ensure_no_port_binding_failure(port) [ 670.437789] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.437789] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] raise exception.PortBindingFailed(port_id=port['id']) [ 670.437789] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] nova.exception.PortBindingFailed: Binding failed for port efa75200-e931-43ca-b12e-4fb50fd45b58, please check neutron logs for more information. [ 670.437789] env[61906]: ERROR nova.compute.manager [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] [ 670.437789] env[61906]: DEBUG nova.compute.utils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Binding failed for port efa75200-e931-43ca-b12e-4fb50fd45b58, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 670.438984] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.458s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.440467] env[61906]: INFO nova.compute.claims [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.443046] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Build of instance e8a2267e-ae4e-4f73-9691-3163ba3ddaa7 was re-scheduled: Binding failed for port efa75200-e931-43ca-b12e-4fb50fd45b58, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 670.443527] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 670.443755] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "refresh_cache-e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.443900] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquired lock "refresh_cache-e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.444083] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 670.580811] env[61906]: DEBUG nova.network.neutron [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.698792] env[61906]: INFO nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: 7aee8304-8d07-4aa2-867a-2eeeeacbabc8] Took 1.02 seconds to deallocate network for instance. [ 670.708658] env[61906]: DEBUG nova.network.neutron [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.975363] env[61906]: DEBUG nova.compute.manager [req-7efe74b1-4520-4ad2-bea7-c8baf2d8bd5f req-7fc4bce1-7d33-4a31-b592-3445856b35d8 service nova] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Received event network-vif-deleted-f1de37fc-eff3-425a-a263-b13b723e3323 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 670.979628] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.131151] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.211042] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Releasing lock "refresh_cache-95604a75-0ab3-45d9-a235-9067e615e72a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.211740] env[61906]: DEBUG nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 671.211991] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 671.212442] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-873abeb1-6cc6-47f5-b438-3a75064150ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.222887] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c00243f4-4edc-4ee3-80a9-0fb294f7d06c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.248788] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 95604a75-0ab3-45d9-a235-9067e615e72a could not be found. [ 671.249184] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 671.249418] env[61906]: INFO nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 671.249666] env[61906]: DEBUG oslo.service.loopingcall [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.249894] env[61906]: DEBUG nova.compute.manager [-] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.249966] env[61906]: DEBUG nova.network.neutron [-] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 671.275571] env[61906]: DEBUG nova.network.neutron [-] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.637624] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Releasing lock "refresh_cache-e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.637624] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 671.637624] env[61906]: DEBUG nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.637624] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 671.655259] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.781296] env[61906]: DEBUG nova.network.neutron [-] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.803988] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d577e19-d873-426a-ad1c-e6ee47f07955 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.813576] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98a70377-807e-492b-bb2b-df3a758352d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.842991] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b01560-9ac5-4630-b6a5-24b524f69080 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.850666] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd5c576-722f-4d4e-ae73-63f8e5fcceba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.865103] env[61906]: DEBUG nova.compute.provider_tree [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.161915] env[61906]: DEBUG nova.network.neutron [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.284176] env[61906]: INFO nova.compute.manager [-] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Took 1.03 seconds to deallocate network for instance. [ 672.286715] env[61906]: DEBUG nova.compute.claims [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 672.286942] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.369068] env[61906]: DEBUG nova.scheduler.client.report [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.662644] env[61906]: INFO nova.compute.manager [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Took 1.03 seconds to deallocate network for instance. [ 672.735174] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "7aee8304-8d07-4aa2-867a-2eeeeacbabc8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.129s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.736448] env[61906]: Traceback (most recent call last): [ 672.736448] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.736448] env[61906]: self.driver.spawn(context, instance, image_meta, [ 672.736448] env[61906]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 672.736448] env[61906]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.736448] env[61906]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.736448] env[61906]: vm_ref = self.build_virtual_machine(instance, [ 672.736448] env[61906]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.736448] env[61906]: vif_infos = vmwarevif.get_vif_info(self._session, [ 672.736448] env[61906]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.736448] env[61906]: for vif in network_info: [ 672.736448] env[61906]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.736448] env[61906]: return self._sync_wrapper(fn, *args, **kwargs) [ 672.736448] env[61906]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.736448] env[61906]: self.wait() [ 672.736448] env[61906]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.736448] env[61906]: self[:] = self._gt.wait() [ 672.736448] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.736448] env[61906]: return self._exit_event.wait() [ 672.736448] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 672.736448] env[61906]: current.throw(*self._exc) [ 672.736448] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.736448] env[61906]: result = function(*args, **kwargs) [ 672.736448] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 672.736448] env[61906]: return func(*args, **kwargs) [ 672.736448] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.736448] env[61906]: raise e [ 672.736448] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.736448] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 672.736448] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.736448] env[61906]: created_port_ids = self._update_ports_for_instance( [ 672.736448] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.736448] env[61906]: with excutils.save_and_reraise_exception(): [ 672.737373] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.737373] env[61906]: self.force_reraise() [ 672.737373] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.737373] env[61906]: raise self.value [ 672.737373] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.737373] env[61906]: updated_port = self._update_port( [ 672.737373] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.737373] env[61906]: _ensure_no_port_binding_failure(port) [ 672.737373] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.737373] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 672.737373] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 7df09b9b-9912-4b71-9cd2-6e0183ee2bde, please check neutron logs for more information. [ 672.737373] env[61906]: During handling of the above exception, another exception occurred: [ 672.737373] env[61906]: Traceback (most recent call last): [ 672.737373] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2456, in _do_build_and_run_instance [ 672.737373] env[61906]: self._build_and_run_instance(context, instance, image, [ 672.737373] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2748, in _build_and_run_instance [ 672.737373] env[61906]: raise exception.RescheduledException( [ 672.737373] env[61906]: nova.exception.RescheduledException: Build of instance 7aee8304-8d07-4aa2-867a-2eeeeacbabc8 was re-scheduled: Binding failed for port 7df09b9b-9912-4b71-9cd2-6e0183ee2bde, please check neutron logs for more information. [ 672.737373] env[61906]: During handling of the above exception, another exception occurred: [ 672.737373] env[61906]: Traceback (most recent call last): [ 672.737373] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 672.737373] env[61906]: func(*args, **kwargs) [ 672.737373] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 672.737373] env[61906]: return func(*args, **kwargs) [ 672.737373] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 672.737373] env[61906]: return f(*args, **kwargs) [ 672.737373] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2347, in _locked_do_build_and_run_instance [ 672.737373] env[61906]: result = self._do_build_and_run_instance(*args, **kwargs) [ 672.738388] env[61906]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 672.738388] env[61906]: with excutils.save_and_reraise_exception(): [ 672.738388] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.738388] env[61906]: self.force_reraise() [ 672.738388] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.738388] env[61906]: raise self.value [ 672.738388] env[61906]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 672.738388] env[61906]: return f(self, context, *args, **kw) [ 672.738388] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 672.738388] env[61906]: with excutils.save_and_reraise_exception(): [ 672.738388] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.738388] env[61906]: self.force_reraise() [ 672.738388] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.738388] env[61906]: raise self.value [ 672.738388] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 672.738388] env[61906]: return function(self, context, *args, **kwargs) [ 672.738388] env[61906]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 672.738388] env[61906]: return function(self, context, *args, **kwargs) [ 672.738388] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 672.738388] env[61906]: return function(self, context, *args, **kwargs) [ 672.738388] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2491, in _do_build_and_run_instance [ 672.738388] env[61906]: instance.save() [ 672.738388] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 672.738388] env[61906]: updates, result = self.indirection_api.object_action( [ 672.738388] env[61906]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 672.738388] env[61906]: return cctxt.call(context, 'object_action', objinst=objinst, [ 672.738388] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 672.738388] env[61906]: result = self.transport._send( [ 672.738388] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 672.738388] env[61906]: return self._driver.send(target, ctxt, message, [ 672.739343] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 672.739343] env[61906]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 672.739343] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 672.739343] env[61906]: raise result [ 672.739343] env[61906]: nova.exception_Remote.InstanceNotFound_Remote: Instance 7aee8304-8d07-4aa2-867a-2eeeeacbabc8 could not be found. [ 672.739343] env[61906]: Traceback (most recent call last): [ 672.739343] env[61906]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 672.739343] env[61906]: return getattr(target, method)(*args, **kwargs) [ 672.739343] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 672.739343] env[61906]: return fn(self, *args, **kwargs) [ 672.739343] env[61906]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 672.739343] env[61906]: old_ref, inst_ref = db.instance_update_and_get_original( [ 672.739343] env[61906]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 672.739343] env[61906]: return f(*args, **kwargs) [ 672.739343] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 672.739343] env[61906]: with excutils.save_and_reraise_exception() as ectxt: [ 672.739343] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.739343] env[61906]: self.force_reraise() [ 672.739343] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.739343] env[61906]: raise self.value [ 672.739343] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 672.739343] env[61906]: return f(*args, **kwargs) [ 672.739343] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 672.739343] env[61906]: return f(context, *args, **kwargs) [ 672.739343] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 672.739343] env[61906]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 672.739343] env[61906]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 672.739343] env[61906]: raise exception.InstanceNotFound(instance_id=uuid) [ 672.739343] env[61906]: nova.exception.InstanceNotFound: Instance 7aee8304-8d07-4aa2-867a-2eeeeacbabc8 could not be found. [ 672.874087] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.874639] env[61906]: DEBUG nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 672.877123] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.499s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.878567] env[61906]: INFO nova.compute.claims [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 672.904238] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Acquiring lock "db4eb67a-12dc-4347-ba29-1af46626a87d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.904497] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Lock "db4eb67a-12dc-4347-ba29-1af46626a87d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.238966] env[61906]: DEBUG nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 673.382777] env[61906]: DEBUG nova.compute.utils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 673.386377] env[61906]: DEBUG nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 673.386641] env[61906]: DEBUG nova.network.neutron [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 673.435632] env[61906]: DEBUG nova.policy [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '217c5f7019504ef8a4d4b11fdbb2a343', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d5318e962dd4c17af474c9d0786a414', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 673.699769] env[61906]: INFO nova.scheduler.client.report [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Deleted allocations for instance e8a2267e-ae4e-4f73-9691-3163ba3ddaa7 [ 673.752896] env[61906]: DEBUG nova.network.neutron [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Successfully created port: 82b16bf3-088a-4900-a616-c820e33458dd {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 673.762223] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.887647] env[61906]: DEBUG nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 674.214133] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ffeb3356-042d-4a7e-be42-daaeb8bacbc7 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.574s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.215555] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.587s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.215780] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "e8a2267e-ae4e-4f73-9691-3163ba3ddaa7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.215982] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "e8a2267e-ae4e-4f73-9691-3163ba3ddaa7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.216159] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "e8a2267e-ae4e-4f73-9691-3163ba3ddaa7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.217841] env[61906]: INFO nova.compute.manager [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Terminating instance [ 674.219314] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquiring lock "refresh_cache-e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.219468] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Acquired lock "refresh_cache-e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.220594] env[61906]: DEBUG nova.network.neutron [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 674.243383] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1612d54-bb27-493f-9ab3-c25585b14f71 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.251814] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9252b2be-cfeb-4368-88d8-cbad6fa89b84 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.283474] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a112b7ce-c7c2-4fa5-801b-323ad8e50454 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.290852] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b06f39-3bd0-4041-8e91-3d659be6b816 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.304841] env[61906]: DEBUG nova.compute.provider_tree [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.568631] env[61906]: DEBUG nova.compute.manager [req-55bc77a5-e7ac-492c-8107-7d61b70ea661 req-68468c64-10d3-41e0-bad6-11b51812f2a3 service nova] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Received event network-changed-82b16bf3-088a-4900-a616-c820e33458dd {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 674.568844] env[61906]: DEBUG nova.compute.manager [req-55bc77a5-e7ac-492c-8107-7d61b70ea661 req-68468c64-10d3-41e0-bad6-11b51812f2a3 service nova] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Refreshing instance network info cache due to event network-changed-82b16bf3-088a-4900-a616-c820e33458dd. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 674.569072] env[61906]: DEBUG oslo_concurrency.lockutils [req-55bc77a5-e7ac-492c-8107-7d61b70ea661 req-68468c64-10d3-41e0-bad6-11b51812f2a3 service nova] Acquiring lock "refresh_cache-42135fac-aef7-47d9-9afe-5ac2351bb07c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.569212] env[61906]: DEBUG oslo_concurrency.lockutils [req-55bc77a5-e7ac-492c-8107-7d61b70ea661 req-68468c64-10d3-41e0-bad6-11b51812f2a3 service nova] Acquired lock "refresh_cache-42135fac-aef7-47d9-9afe-5ac2351bb07c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.569368] env[61906]: DEBUG nova.network.neutron [req-55bc77a5-e7ac-492c-8107-7d61b70ea661 req-68468c64-10d3-41e0-bad6-11b51812f2a3 service nova] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Refreshing network info cache for port 82b16bf3-088a-4900-a616-c820e33458dd {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 674.718404] env[61906]: DEBUG nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 674.739861] env[61906]: DEBUG nova.network.neutron [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.807626] env[61906]: DEBUG nova.scheduler.client.report [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.881556] env[61906]: DEBUG nova.network.neutron [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.901400] env[61906]: DEBUG nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 674.930300] env[61906]: DEBUG nova.virt.hardware [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 674.930300] env[61906]: DEBUG nova.virt.hardware [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 674.930300] env[61906]: DEBUG nova.virt.hardware [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 674.930893] env[61906]: DEBUG nova.virt.hardware [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 674.930893] env[61906]: DEBUG nova.virt.hardware [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 674.930893] env[61906]: DEBUG nova.virt.hardware [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 674.930893] env[61906]: DEBUG nova.virt.hardware [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 674.930893] env[61906]: DEBUG nova.virt.hardware [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 674.931155] env[61906]: DEBUG nova.virt.hardware [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 674.931155] env[61906]: DEBUG nova.virt.hardware [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 674.931155] env[61906]: DEBUG nova.virt.hardware [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 674.931352] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f10147-2577-464b-824a-176ca74a95a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.934843] env[61906]: ERROR nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 82b16bf3-088a-4900-a616-c820e33458dd, please check neutron logs for more information. [ 674.934843] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 674.934843] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.934843] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 674.934843] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.934843] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 674.934843] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.934843] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 674.934843] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.934843] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 674.934843] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.934843] env[61906]: ERROR nova.compute.manager raise self.value [ 674.934843] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.934843] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 674.934843] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.934843] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 674.935684] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.935684] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 674.935684] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 82b16bf3-088a-4900-a616-c820e33458dd, please check neutron logs for more information. [ 674.935684] env[61906]: ERROR nova.compute.manager [ 674.935684] env[61906]: Traceback (most recent call last): [ 674.935684] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 674.935684] env[61906]: listener.cb(fileno) [ 674.935684] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.935684] env[61906]: result = function(*args, **kwargs) [ 674.935684] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 674.935684] env[61906]: return func(*args, **kwargs) [ 674.935684] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.935684] env[61906]: raise e [ 674.935684] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.935684] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 674.935684] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.935684] env[61906]: created_port_ids = self._update_ports_for_instance( [ 674.935684] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.935684] env[61906]: with excutils.save_and_reraise_exception(): [ 674.935684] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.935684] env[61906]: self.force_reraise() [ 674.935684] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.935684] env[61906]: raise self.value [ 674.935684] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.935684] env[61906]: updated_port = self._update_port( [ 674.935684] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.935684] env[61906]: _ensure_no_port_binding_failure(port) [ 674.935684] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.935684] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 674.936987] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 82b16bf3-088a-4900-a616-c820e33458dd, please check neutron logs for more information. [ 674.936987] env[61906]: Removing descriptor: 17 [ 674.940981] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e8113e-7889-42d3-90a9-6bf77d051803 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.955952] env[61906]: ERROR nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 82b16bf3-088a-4900-a616-c820e33458dd, please check neutron logs for more information. [ 674.955952] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Traceback (most recent call last): [ 674.955952] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 674.955952] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] yield resources [ 674.955952] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.955952] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] self.driver.spawn(context, instance, image_meta, [ 674.955952] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 674.955952] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.955952] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.955952] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] vm_ref = self.build_virtual_machine(instance, [ 674.955952] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.956601] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.956601] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.956601] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] for vif in network_info: [ 674.956601] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.956601] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] return self._sync_wrapper(fn, *args, **kwargs) [ 674.956601] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.956601] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] self.wait() [ 674.956601] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.956601] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] self[:] = self._gt.wait() [ 674.956601] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.956601] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] return self._exit_event.wait() [ 674.956601] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 674.956601] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] current.throw(*self._exc) [ 674.957251] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.957251] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] result = function(*args, **kwargs) [ 674.957251] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 674.957251] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] return func(*args, **kwargs) [ 674.957251] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.957251] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] raise e [ 674.957251] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.957251] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] nwinfo = self.network_api.allocate_for_instance( [ 674.957251] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.957251] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] created_port_ids = self._update_ports_for_instance( [ 674.957251] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.957251] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] with excutils.save_and_reraise_exception(): [ 674.957251] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.957861] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] self.force_reraise() [ 674.957861] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.957861] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] raise self.value [ 674.957861] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.957861] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] updated_port = self._update_port( [ 674.957861] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.957861] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] _ensure_no_port_binding_failure(port) [ 674.957861] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.957861] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] raise exception.PortBindingFailed(port_id=port['id']) [ 674.957861] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] nova.exception.PortBindingFailed: Binding failed for port 82b16bf3-088a-4900-a616-c820e33458dd, please check neutron logs for more information. [ 674.957861] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] [ 674.957861] env[61906]: INFO nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Terminating instance [ 674.958439] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "refresh_cache-42135fac-aef7-47d9-9afe-5ac2351bb07c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.093476] env[61906]: DEBUG nova.network.neutron [req-55bc77a5-e7ac-492c-8107-7d61b70ea661 req-68468c64-10d3-41e0-bad6-11b51812f2a3 service nova] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.203672] env[61906]: DEBUG nova.network.neutron [req-55bc77a5-e7ac-492c-8107-7d61b70ea661 req-68468c64-10d3-41e0-bad6-11b51812f2a3 service nova] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.247055] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.312534] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.313119] env[61906]: DEBUG nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 675.316404] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.558s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.317861] env[61906]: INFO nova.compute.claims [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 675.384586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Releasing lock "refresh_cache-e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.385014] env[61906]: DEBUG nova.compute.manager [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 675.385210] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 675.386074] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f4d8a9a-bfbc-4c34-9e18-565d034114c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.400382] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ecd9b5-4a8a-46ff-a6b5-e009939663aa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.420950] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e8a2267e-ae4e-4f73-9691-3163ba3ddaa7 could not be found. [ 675.421500] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 675.421500] env[61906]: INFO nova.compute.manager [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 675.421634] env[61906]: DEBUG oslo.service.loopingcall [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 675.421740] env[61906]: DEBUG nova.compute.manager [-] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 675.422790] env[61906]: DEBUG nova.network.neutron [-] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 675.445079] env[61906]: DEBUG nova.network.neutron [-] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.707773] env[61906]: DEBUG oslo_concurrency.lockutils [req-55bc77a5-e7ac-492c-8107-7d61b70ea661 req-68468c64-10d3-41e0-bad6-11b51812f2a3 service nova] Releasing lock "refresh_cache-42135fac-aef7-47d9-9afe-5ac2351bb07c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.707773] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquired lock "refresh_cache-42135fac-aef7-47d9-9afe-5ac2351bb07c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.707773] env[61906]: DEBUG nova.network.neutron [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.823791] env[61906]: DEBUG nova.compute.utils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 675.826626] env[61906]: DEBUG nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 675.827149] env[61906]: DEBUG nova.network.neutron [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 675.889280] env[61906]: DEBUG nova.policy [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98ce5f6913eb45c58762ff5e67bfb4ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6c3dc3a6a9e34ec88019bb84edfa6cad', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 675.947926] env[61906]: DEBUG nova.network.neutron [-] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.183854] env[61906]: DEBUG nova.network.neutron [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Successfully created port: 854dca8c-7653-47d3-a584-6fd29de10059 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.235692] env[61906]: DEBUG nova.network.neutron [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.327357] env[61906]: DEBUG nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 676.380754] env[61906]: DEBUG nova.network.neutron [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.453176] env[61906]: INFO nova.compute.manager [-] [instance: e8a2267e-ae4e-4f73-9691-3163ba3ddaa7] Took 1.03 seconds to deallocate network for instance. [ 676.607126] env[61906]: DEBUG nova.compute.manager [req-7688c9b9-e0ee-4877-ac3c-f77896eea811 req-ac2c9e10-adc4-4339-81a2-826a8fb41a92 service nova] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Received event network-vif-deleted-82b16bf3-088a-4900-a616-c820e33458dd {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 676.711948] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb13b84-38b2-4112-a18a-9f1e7ffe9106 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.719892] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d94da2a1-1edb-4f4a-92d5-f34d313ecb63 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.752157] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7855bd02-6292-4afe-85e1-1fd08a755f34 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.759833] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea491546-765b-424a-8126-892a628281a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.774603] env[61906]: DEBUG nova.compute.provider_tree [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.886069] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Releasing lock "refresh_cache-42135fac-aef7-47d9-9afe-5ac2351bb07c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.886069] env[61906]: DEBUG nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 676.886069] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 676.886069] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee216998-c009-494a-ad63-5f239178bad9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.896115] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62bcc46b-03d5-4a8a-beb6-9e365c1bde23 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.916800] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 42135fac-aef7-47d9-9afe-5ac2351bb07c could not be found. [ 676.917040] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 676.917226] env[61906]: INFO nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 676.917484] env[61906]: DEBUG oslo.service.loopingcall [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 676.917690] env[61906]: DEBUG nova.compute.manager [-] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.917782] env[61906]: DEBUG nova.network.neutron [-] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.943409] env[61906]: DEBUG nova.network.neutron [-] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.277846] env[61906]: DEBUG nova.scheduler.client.report [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.289675] env[61906]: ERROR nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 854dca8c-7653-47d3-a584-6fd29de10059, please check neutron logs for more information. [ 677.289675] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 677.289675] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.289675] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 677.289675] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.289675] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 677.289675] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.289675] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 677.289675] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.289675] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 677.289675] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.289675] env[61906]: ERROR nova.compute.manager raise self.value [ 677.289675] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.289675] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 677.289675] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.289675] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 677.290226] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.290226] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 677.290226] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 854dca8c-7653-47d3-a584-6fd29de10059, please check neutron logs for more information. [ 677.290226] env[61906]: ERROR nova.compute.manager [ 677.290226] env[61906]: Traceback (most recent call last): [ 677.290226] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 677.290226] env[61906]: listener.cb(fileno) [ 677.290226] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.290226] env[61906]: result = function(*args, **kwargs) [ 677.290226] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 677.290226] env[61906]: return func(*args, **kwargs) [ 677.290226] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.290226] env[61906]: raise e [ 677.290226] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.290226] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 677.290226] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.290226] env[61906]: created_port_ids = self._update_ports_for_instance( [ 677.290226] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.290226] env[61906]: with excutils.save_and_reraise_exception(): [ 677.290226] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.290226] env[61906]: self.force_reraise() [ 677.290226] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.290226] env[61906]: raise self.value [ 677.290226] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.290226] env[61906]: updated_port = self._update_port( [ 677.290226] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.290226] env[61906]: _ensure_no_port_binding_failure(port) [ 677.290226] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.290226] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 677.291048] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 854dca8c-7653-47d3-a584-6fd29de10059, please check neutron logs for more information. [ 677.291048] env[61906]: Removing descriptor: 17 [ 677.338784] env[61906]: DEBUG nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.361470] env[61906]: DEBUG nova.virt.hardware [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.361714] env[61906]: DEBUG nova.virt.hardware [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.361868] env[61906]: DEBUG nova.virt.hardware [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.362060] env[61906]: DEBUG nova.virt.hardware [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.362211] env[61906]: DEBUG nova.virt.hardware [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.362358] env[61906]: DEBUG nova.virt.hardware [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.362562] env[61906]: DEBUG nova.virt.hardware [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.362715] env[61906]: DEBUG nova.virt.hardware [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.362876] env[61906]: DEBUG nova.virt.hardware [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.363043] env[61906]: DEBUG nova.virt.hardware [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.363215] env[61906]: DEBUG nova.virt.hardware [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.364117] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef9c49f0-d10e-41c5-91b7-0d75d70f0d32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.372786] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-224c510a-48f9-4c63-a901-4887858dc54f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.386693] env[61906]: ERROR nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 854dca8c-7653-47d3-a584-6fd29de10059, please check neutron logs for more information. [ 677.386693] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Traceback (most recent call last): [ 677.386693] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 677.386693] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] yield resources [ 677.386693] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.386693] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] self.driver.spawn(context, instance, image_meta, [ 677.386693] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 677.386693] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.386693] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.386693] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] vm_ref = self.build_virtual_machine(instance, [ 677.386693] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.387087] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.387087] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.387087] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] for vif in network_info: [ 677.387087] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.387087] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] return self._sync_wrapper(fn, *args, **kwargs) [ 677.387087] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.387087] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] self.wait() [ 677.387087] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.387087] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] self[:] = self._gt.wait() [ 677.387087] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.387087] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] return self._exit_event.wait() [ 677.387087] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 677.387087] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] current.throw(*self._exc) [ 677.387474] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.387474] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] result = function(*args, **kwargs) [ 677.387474] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 677.387474] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] return func(*args, **kwargs) [ 677.387474] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.387474] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] raise e [ 677.387474] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.387474] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] nwinfo = self.network_api.allocate_for_instance( [ 677.387474] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.387474] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] created_port_ids = self._update_ports_for_instance( [ 677.387474] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.387474] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] with excutils.save_and_reraise_exception(): [ 677.387474] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.387917] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] self.force_reraise() [ 677.387917] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.387917] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] raise self.value [ 677.387917] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.387917] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] updated_port = self._update_port( [ 677.387917] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.387917] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] _ensure_no_port_binding_failure(port) [ 677.387917] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.387917] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] raise exception.PortBindingFailed(port_id=port['id']) [ 677.387917] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] nova.exception.PortBindingFailed: Binding failed for port 854dca8c-7653-47d3-a584-6fd29de10059, please check neutron logs for more information. [ 677.387917] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] [ 677.387917] env[61906]: INFO nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Terminating instance [ 677.389007] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Acquiring lock "refresh_cache-71fe23a8-1108-4781-9a35-0156cb9d584d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.389179] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Acquired lock "refresh_cache-71fe23a8-1108-4781-9a35-0156cb9d584d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.389346] env[61906]: DEBUG nova.network.neutron [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 677.445408] env[61906]: DEBUG nova.network.neutron [-] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.482912] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b2a72d99-417a-4840-98f9-f83c46d2cbb8 tempest-ListServersNegativeTestJSON-567814738 tempest-ListServersNegativeTestJSON-567814738-project-member] Lock "e8a2267e-ae4e-4f73-9691-3163ba3ddaa7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.267s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.784131] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.784131] env[61906]: DEBUG nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 677.786245] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.827s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.906975] env[61906]: DEBUG nova.network.neutron [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.947782] env[61906]: INFO nova.compute.manager [-] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Took 1.03 seconds to deallocate network for instance. [ 677.950362] env[61906]: DEBUG nova.compute.claims [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 677.950571] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.995441] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquiring lock "a6e54432-a58e-49ba-a5c8-e6188f2ea4ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.995673] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Lock "a6e54432-a58e-49ba-a5c8-e6188f2ea4ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.002698] env[61906]: DEBUG nova.network.neutron [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.290278] env[61906]: DEBUG nova.compute.utils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 678.294873] env[61906]: DEBUG nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 678.295061] env[61906]: DEBUG nova.network.neutron [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 678.342094] env[61906]: DEBUG nova.policy [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ded7bad07b5e4f31ba627d4e4dbea0d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '473163e61f064b969326e695af5c54b1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 678.509599] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Releasing lock "refresh_cache-71fe23a8-1108-4781-9a35-0156cb9d584d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.510025] env[61906]: DEBUG nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 678.510223] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 678.511200] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2935aec7-f47b-40b5-b019-d47a0557e34f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.521990] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6174e580-7cf6-4d1b-a79b-4a65ffe070e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.545069] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 71fe23a8-1108-4781-9a35-0156cb9d584d could not be found. [ 678.545301] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 678.545481] env[61906]: INFO nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 678.545725] env[61906]: DEBUG oslo.service.loopingcall [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 678.547745] env[61906]: DEBUG nova.compute.manager [-] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.547745] env[61906]: DEBUG nova.network.neutron [-] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 678.567821] env[61906]: DEBUG nova.network.neutron [-] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.626838] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97494175-5e1e-447f-9216-d745cb1118df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.636052] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d278506f-925d-4648-9342-78d8c47dbc90 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.666284] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9685607c-3f66-47c1-963d-0c6545ae7c8f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.669728] env[61906]: DEBUG nova.compute.manager [req-feebfd46-60c7-4853-8e86-91cbbed36cdf req-a934753b-fa14-41bb-b199-03b25e542ae3 service nova] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Received event network-changed-854dca8c-7653-47d3-a584-6fd29de10059 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 678.669918] env[61906]: DEBUG nova.compute.manager [req-feebfd46-60c7-4853-8e86-91cbbed36cdf req-a934753b-fa14-41bb-b199-03b25e542ae3 service nova] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Refreshing instance network info cache due to event network-changed-854dca8c-7653-47d3-a584-6fd29de10059. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 678.670155] env[61906]: DEBUG oslo_concurrency.lockutils [req-feebfd46-60c7-4853-8e86-91cbbed36cdf req-a934753b-fa14-41bb-b199-03b25e542ae3 service nova] Acquiring lock "refresh_cache-71fe23a8-1108-4781-9a35-0156cb9d584d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.670341] env[61906]: DEBUG oslo_concurrency.lockutils [req-feebfd46-60c7-4853-8e86-91cbbed36cdf req-a934753b-fa14-41bb-b199-03b25e542ae3 service nova] Acquired lock "refresh_cache-71fe23a8-1108-4781-9a35-0156cb9d584d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.670508] env[61906]: DEBUG nova.network.neutron [req-feebfd46-60c7-4853-8e86-91cbbed36cdf req-a934753b-fa14-41bb-b199-03b25e542ae3 service nova] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Refreshing network info cache for port 854dca8c-7653-47d3-a584-6fd29de10059 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 678.677157] env[61906]: DEBUG nova.network.neutron [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Successfully created port: 96e6d7af-92f0-4732-be60-755f37c5d0f3 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 678.679462] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022687db-039b-4280-8cb1-071f8ff4e847 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.694126] env[61906]: DEBUG nova.compute.provider_tree [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.795798] env[61906]: DEBUG nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 678.957016] env[61906]: DEBUG nova.network.neutron [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Successfully created port: 687133b6-f678-47f5-88ab-d22c62dc30cc {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 679.070940] env[61906]: DEBUG nova.network.neutron [-] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.189619] env[61906]: DEBUG nova.network.neutron [req-feebfd46-60c7-4853-8e86-91cbbed36cdf req-a934753b-fa14-41bb-b199-03b25e542ae3 service nova] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.197392] env[61906]: DEBUG nova.scheduler.client.report [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.219487] env[61906]: DEBUG nova.network.neutron [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Successfully created port: 76aaa92b-cae2-43ed-bc2e-2430a439c5f9 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 679.260923] env[61906]: DEBUG nova.network.neutron [req-feebfd46-60c7-4853-8e86-91cbbed36cdf req-a934753b-fa14-41bb-b199-03b25e542ae3 service nova] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.573916] env[61906]: INFO nova.compute.manager [-] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Took 1.03 seconds to deallocate network for instance. [ 679.576387] env[61906]: DEBUG nova.compute.claims [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 679.576565] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.704253] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.917s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.704253] env[61906]: ERROR nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4980c357-94cf-4788-a94c-3d0d0377dd59, please check neutron logs for more information. [ 679.704253] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Traceback (most recent call last): [ 679.704253] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.704253] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] self.driver.spawn(context, instance, image_meta, [ 679.704253] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 679.704253] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.704253] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.704253] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] vm_ref = self.build_virtual_machine(instance, [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] for vif in network_info: [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] return self._sync_wrapper(fn, *args, **kwargs) [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] self.wait() [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] self[:] = self._gt.wait() [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] return self._exit_event.wait() [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 679.704582] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] current.throw(*self._exc) [ 679.705070] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.705070] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] result = function(*args, **kwargs) [ 679.705070] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 679.705070] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] return func(*args, **kwargs) [ 679.705070] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.705070] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] raise e [ 679.705070] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.705070] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] nwinfo = self.network_api.allocate_for_instance( [ 679.705070] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.705070] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] created_port_ids = self._update_ports_for_instance( [ 679.705070] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.705070] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] with excutils.save_and_reraise_exception(): [ 679.705070] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.705467] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] self.force_reraise() [ 679.705467] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.705467] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] raise self.value [ 679.705467] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.705467] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] updated_port = self._update_port( [ 679.705467] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.705467] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] _ensure_no_port_binding_failure(port) [ 679.705467] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.705467] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] raise exception.PortBindingFailed(port_id=port['id']) [ 679.705467] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] nova.exception.PortBindingFailed: Binding failed for port 4980c357-94cf-4788-a94c-3d0d0377dd59, please check neutron logs for more information. [ 679.705467] env[61906]: ERROR nova.compute.manager [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] [ 679.705783] env[61906]: DEBUG nova.compute.utils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Binding failed for port 4980c357-94cf-4788-a94c-3d0d0377dd59, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 679.705821] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.063s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.712863] env[61906]: DEBUG nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Build of instance 569d7ab1-22b8-4006-bf0e-ff35032cb9a6 was re-scheduled: Binding failed for port 4980c357-94cf-4788-a94c-3d0d0377dd59, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 679.714407] env[61906]: DEBUG nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 679.714642] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "refresh_cache-569d7ab1-22b8-4006-bf0e-ff35032cb9a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.714793] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "refresh_cache-569d7ab1-22b8-4006-bf0e-ff35032cb9a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.714953] env[61906]: DEBUG nova.network.neutron [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 679.762978] env[61906]: DEBUG oslo_concurrency.lockutils [req-feebfd46-60c7-4853-8e86-91cbbed36cdf req-a934753b-fa14-41bb-b199-03b25e542ae3 service nova] Releasing lock "refresh_cache-71fe23a8-1108-4781-9a35-0156cb9d584d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.763279] env[61906]: DEBUG nova.compute.manager [req-feebfd46-60c7-4853-8e86-91cbbed36cdf req-a934753b-fa14-41bb-b199-03b25e542ae3 service nova] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Received event network-vif-deleted-854dca8c-7653-47d3-a584-6fd29de10059 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 679.807262] env[61906]: DEBUG nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 679.839094] env[61906]: DEBUG nova.virt.hardware [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 679.839094] env[61906]: DEBUG nova.virt.hardware [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 679.839094] env[61906]: DEBUG nova.virt.hardware [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 679.839094] env[61906]: DEBUG nova.virt.hardware [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 679.839309] env[61906]: DEBUG nova.virt.hardware [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 679.839309] env[61906]: DEBUG nova.virt.hardware [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 679.839309] env[61906]: DEBUG nova.virt.hardware [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 679.839309] env[61906]: DEBUG nova.virt.hardware [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 679.839309] env[61906]: DEBUG nova.virt.hardware [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 679.839577] env[61906]: DEBUG nova.virt.hardware [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 679.839897] env[61906]: DEBUG nova.virt.hardware [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 679.840850] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91e7dd1-1cbe-4299-9333-57cc121115f6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.854915] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd6df59-4438-4710-bb73-679b8b3ca731 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.236147] env[61906]: DEBUG nova.network.neutron [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.239159] env[61906]: ERROR nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 96e6d7af-92f0-4732-be60-755f37c5d0f3, please check neutron logs for more information. [ 680.239159] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 680.239159] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.239159] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 680.239159] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.239159] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 680.239159] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.239159] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 680.239159] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.239159] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 680.239159] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.239159] env[61906]: ERROR nova.compute.manager raise self.value [ 680.239159] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.239159] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 680.239159] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.239159] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 680.239606] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.239606] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 680.239606] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 96e6d7af-92f0-4732-be60-755f37c5d0f3, please check neutron logs for more information. [ 680.239606] env[61906]: ERROR nova.compute.manager [ 680.239606] env[61906]: Traceback (most recent call last): [ 680.239606] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 680.239606] env[61906]: listener.cb(fileno) [ 680.239606] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.239606] env[61906]: result = function(*args, **kwargs) [ 680.239606] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 680.239606] env[61906]: return func(*args, **kwargs) [ 680.239606] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.239606] env[61906]: raise e [ 680.239606] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.239606] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 680.239606] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.239606] env[61906]: created_port_ids = self._update_ports_for_instance( [ 680.239606] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.239606] env[61906]: with excutils.save_and_reraise_exception(): [ 680.239606] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.239606] env[61906]: self.force_reraise() [ 680.239606] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.239606] env[61906]: raise self.value [ 680.239606] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.239606] env[61906]: updated_port = self._update_port( [ 680.239606] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.239606] env[61906]: _ensure_no_port_binding_failure(port) [ 680.239606] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.239606] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 680.240327] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 96e6d7af-92f0-4732-be60-755f37c5d0f3, please check neutron logs for more information. [ 680.240327] env[61906]: Removing descriptor: 17 [ 680.240327] env[61906]: ERROR nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 96e6d7af-92f0-4732-be60-755f37c5d0f3, please check neutron logs for more information. [ 680.240327] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Traceback (most recent call last): [ 680.240327] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 680.240327] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] yield resources [ 680.240327] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.240327] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] self.driver.spawn(context, instance, image_meta, [ 680.240327] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 680.240327] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.240327] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.240327] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] vm_ref = self.build_virtual_machine(instance, [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] for vif in network_info: [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] return self._sync_wrapper(fn, *args, **kwargs) [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] self.wait() [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] self[:] = self._gt.wait() [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] return self._exit_event.wait() [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 680.240711] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] result = hub.switch() [ 680.241135] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 680.241135] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] return self.greenlet.switch() [ 680.241135] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.241135] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] result = function(*args, **kwargs) [ 680.241135] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 680.241135] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] return func(*args, **kwargs) [ 680.241135] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.241135] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] raise e [ 680.241135] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.241135] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] nwinfo = self.network_api.allocate_for_instance( [ 680.241135] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.241135] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] created_port_ids = self._update_ports_for_instance( [ 680.241135] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.241489] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] with excutils.save_and_reraise_exception(): [ 680.241489] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.241489] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] self.force_reraise() [ 680.241489] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.241489] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] raise self.value [ 680.241489] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.241489] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] updated_port = self._update_port( [ 680.241489] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.241489] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] _ensure_no_port_binding_failure(port) [ 680.241489] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.241489] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] raise exception.PortBindingFailed(port_id=port['id']) [ 680.241489] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] nova.exception.PortBindingFailed: Binding failed for port 96e6d7af-92f0-4732-be60-755f37c5d0f3, please check neutron logs for more information. [ 680.241489] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] [ 680.241852] env[61906]: INFO nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Terminating instance [ 680.242216] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquiring lock "refresh_cache-8d4c25cc-47d0-42e0-8fe9-49426175af68" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.242374] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquired lock "refresh_cache-8d4c25cc-47d0-42e0-8fe9-49426175af68" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.242532] env[61906]: DEBUG nova.network.neutron [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 680.313537] env[61906]: DEBUG nova.network.neutron [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.519666] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9634e754-d56f-412c-8f8b-8c6260de8bbf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.528569] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da2ff5c-a9c4-40b1-b9ae-49b9a57a6d9f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.560071] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ea9ed7-a2d5-43ba-a465-aa50635c5e1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.567753] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9218cd-5d01-4caa-83a0-8c80e6d911ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.581873] env[61906]: DEBUG nova.compute.provider_tree [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.690380] env[61906]: DEBUG nova.compute.manager [req-c05b7bb0-7512-4632-992c-636a53a2965b req-2b7c3e39-3cac-4a37-a4f3-fde2526f7880 service nova] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Received event network-changed-96e6d7af-92f0-4732-be60-755f37c5d0f3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 680.690821] env[61906]: DEBUG nova.compute.manager [req-c05b7bb0-7512-4632-992c-636a53a2965b req-2b7c3e39-3cac-4a37-a4f3-fde2526f7880 service nova] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Refreshing instance network info cache due to event network-changed-96e6d7af-92f0-4732-be60-755f37c5d0f3. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 680.691300] env[61906]: DEBUG oslo_concurrency.lockutils [req-c05b7bb0-7512-4632-992c-636a53a2965b req-2b7c3e39-3cac-4a37-a4f3-fde2526f7880 service nova] Acquiring lock "refresh_cache-8d4c25cc-47d0-42e0-8fe9-49426175af68" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.815772] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "refresh_cache-569d7ab1-22b8-4006-bf0e-ff35032cb9a6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.816998] env[61906]: DEBUG nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 680.816998] env[61906]: DEBUG nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.816998] env[61906]: DEBUG nova.network.neutron [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 680.835785] env[61906]: DEBUG nova.network.neutron [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.843468] env[61906]: DEBUG nova.network.neutron [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.944203] env[61906]: DEBUG nova.network.neutron [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.084852] env[61906]: DEBUG nova.scheduler.client.report [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.346905] env[61906]: DEBUG nova.network.neutron [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.448181] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Releasing lock "refresh_cache-8d4c25cc-47d0-42e0-8fe9-49426175af68" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.448595] env[61906]: DEBUG nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 681.448791] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 681.449128] env[61906]: DEBUG oslo_concurrency.lockutils [req-c05b7bb0-7512-4632-992c-636a53a2965b req-2b7c3e39-3cac-4a37-a4f3-fde2526f7880 service nova] Acquired lock "refresh_cache-8d4c25cc-47d0-42e0-8fe9-49426175af68" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.449300] env[61906]: DEBUG nova.network.neutron [req-c05b7bb0-7512-4632-992c-636a53a2965b req-2b7c3e39-3cac-4a37-a4f3-fde2526f7880 service nova] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Refreshing network info cache for port 96e6d7af-92f0-4732-be60-755f37c5d0f3 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 681.450402] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-75bb6635-6524-4d0e-9528-610ace5ab01d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.461063] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c13a4b-8b26-41a6-af04-9dcbba1053ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.482811] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8d4c25cc-47d0-42e0-8fe9-49426175af68 could not be found. [ 681.483066] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 681.483258] env[61906]: INFO nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Took 0.03 seconds to destroy the instance on the hypervisor. [ 681.483626] env[61906]: DEBUG oslo.service.loopingcall [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 681.483772] env[61906]: DEBUG nova.compute.manager [-] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.483863] env[61906]: DEBUG nova.network.neutron [-] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 681.582188] env[61906]: DEBUG nova.network.neutron [-] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.589832] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.884s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.590468] env[61906]: ERROR nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 16f2305d-3260-4121-9d5c-89b1cf66ccf0, please check neutron logs for more information. [ 681.590468] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Traceback (most recent call last): [ 681.590468] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.590468] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] self.driver.spawn(context, instance, image_meta, [ 681.590468] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 681.590468] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.590468] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.590468] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] vm_ref = self.build_virtual_machine(instance, [ 681.590468] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.590468] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.590468] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] for vif in network_info: [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] return self._sync_wrapper(fn, *args, **kwargs) [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] self.wait() [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] self[:] = self._gt.wait() [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] return self._exit_event.wait() [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] current.throw(*self._exc) [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] result = function(*args, **kwargs) [ 681.590885] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 681.591261] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] return func(*args, **kwargs) [ 681.591261] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.591261] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] raise e [ 681.591261] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.591261] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] nwinfo = self.network_api.allocate_for_instance( [ 681.591261] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.591261] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] created_port_ids = self._update_ports_for_instance( [ 681.591261] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.591261] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] with excutils.save_and_reraise_exception(): [ 681.591261] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.591261] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] self.force_reraise() [ 681.591261] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.591261] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] raise self.value [ 681.591584] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.591584] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] updated_port = self._update_port( [ 681.591584] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.591584] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] _ensure_no_port_binding_failure(port) [ 681.591584] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.591584] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] raise exception.PortBindingFailed(port_id=port['id']) [ 681.591584] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] nova.exception.PortBindingFailed: Binding failed for port 16f2305d-3260-4121-9d5c-89b1cf66ccf0, please check neutron logs for more information. [ 681.591584] env[61906]: ERROR nova.compute.manager [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] [ 681.591584] env[61906]: DEBUG nova.compute.utils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Binding failed for port 16f2305d-3260-4121-9d5c-89b1cf66ccf0, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 681.592299] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.555s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.593896] env[61906]: INFO nova.compute.claims [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 681.596465] env[61906]: DEBUG nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Build of instance e5c81f8b-aac3-4cf8-9abb-444731f57206 was re-scheduled: Binding failed for port 16f2305d-3260-4121-9d5c-89b1cf66ccf0, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 681.596882] env[61906]: DEBUG nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 681.597112] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquiring lock "refresh_cache-e5c81f8b-aac3-4cf8-9abb-444731f57206" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.597260] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquired lock "refresh_cache-e5c81f8b-aac3-4cf8-9abb-444731f57206" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.597415] env[61906]: DEBUG nova.network.neutron [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 681.849794] env[61906]: INFO nova.compute.manager [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 569d7ab1-22b8-4006-bf0e-ff35032cb9a6] Took 1.03 seconds to deallocate network for instance. [ 682.004466] env[61906]: DEBUG nova.network.neutron [req-c05b7bb0-7512-4632-992c-636a53a2965b req-2b7c3e39-3cac-4a37-a4f3-fde2526f7880 service nova] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.113725] env[61906]: DEBUG nova.network.neutron [req-c05b7bb0-7512-4632-992c-636a53a2965b req-2b7c3e39-3cac-4a37-a4f3-fde2526f7880 service nova] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.126334] env[61906]: DEBUG nova.network.neutron [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.231126] env[61906]: DEBUG nova.network.neutron [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.545189] env[61906]: DEBUG nova.network.neutron [-] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.616691] env[61906]: DEBUG oslo_concurrency.lockutils [req-c05b7bb0-7512-4632-992c-636a53a2965b req-2b7c3e39-3cac-4a37-a4f3-fde2526f7880 service nova] Releasing lock "refresh_cache-8d4c25cc-47d0-42e0-8fe9-49426175af68" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.616956] env[61906]: DEBUG nova.compute.manager [req-c05b7bb0-7512-4632-992c-636a53a2965b req-2b7c3e39-3cac-4a37-a4f3-fde2526f7880 service nova] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Received event network-vif-deleted-96e6d7af-92f0-4732-be60-755f37c5d0f3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 682.733478] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Releasing lock "refresh_cache-e5c81f8b-aac3-4cf8-9abb-444731f57206" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.733727] env[61906]: DEBUG nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 682.733910] env[61906]: DEBUG nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.734120] env[61906]: DEBUG nova.network.neutron [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 682.750636] env[61906]: DEBUG nova.network.neutron [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.877440] env[61906]: INFO nova.scheduler.client.report [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Deleted allocations for instance 569d7ab1-22b8-4006-bf0e-ff35032cb9a6 [ 682.886667] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1b3b37-3c07-4542-8465-7ece92675981 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.894845] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990afbed-96b4-40f2-801c-f0eaebd81b46 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.925504] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d1276b-2ce9-441e-b15e-afcf88ae6d65 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.933470] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31c56e0-6398-46d8-b595-c3fbf512b012 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.948240] env[61906]: DEBUG nova.compute.provider_tree [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.047164] env[61906]: INFO nova.compute.manager [-] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Took 1.56 seconds to deallocate network for instance. [ 683.049731] env[61906]: DEBUG nova.compute.claims [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 683.049731] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.252597] env[61906]: DEBUG nova.network.neutron [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.390057] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8fad782f-c60d-4960-9378-073d1f81fdee tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "569d7ab1-22b8-4006-bf0e-ff35032cb9a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.241s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.451068] env[61906]: DEBUG nova.scheduler.client.report [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.757224] env[61906]: INFO nova.compute.manager [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: e5c81f8b-aac3-4cf8-9abb-444731f57206] Took 1.02 seconds to deallocate network for instance. [ 683.893077] env[61906]: DEBUG nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 683.955795] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.956334] env[61906]: DEBUG nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 683.959074] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.035s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.960514] env[61906]: INFO nova.compute.claims [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.415028] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.465161] env[61906]: DEBUG nova.compute.utils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 684.468420] env[61906]: DEBUG nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 684.468591] env[61906]: DEBUG nova.network.neutron [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 684.534492] env[61906]: DEBUG nova.policy [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ffa49615fdd94b3abfb1a73d5de6dcba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75cc323f1ae54f268d4c515895be0065', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 684.792185] env[61906]: INFO nova.scheduler.client.report [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Deleted allocations for instance e5c81f8b-aac3-4cf8-9abb-444731f57206 [ 684.870798] env[61906]: DEBUG nova.network.neutron [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Successfully created port: 3dae35ba-bed5-44cb-a046-47bcc1cdc551 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 684.971752] env[61906]: DEBUG nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 685.300531] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6f3625b-b38a-47d7-b4e1-9592677b92c8 tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Lock "e5c81f8b-aac3-4cf8-9abb-444731f57206" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.093s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.341163] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81716094-e2b7-487e-9e8c-dc6cda3002c9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.349052] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8185d033-b8b3-4357-bc77-d0fe10d3bcf7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.382746] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-febd1910-52fd-462e-8e81-617d84939003 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.391975] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d6f06d-7f21-46d9-93aa-652d765b4df3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.398933] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "b1429121-3c7e-4982-9059-5eb98a7f92df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.399177] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "b1429121-3c7e-4982-9059-5eb98a7f92df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.413043] env[61906]: DEBUG nova.compute.provider_tree [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.666242] env[61906]: DEBUG nova.compute.manager [req-5b9803bb-2a46-4a74-bc44-d82282703f9f req-672af96c-d32f-422f-a93f-1e75e0a77d55 service nova] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Received event network-changed-3dae35ba-bed5-44cb-a046-47bcc1cdc551 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 685.666242] env[61906]: DEBUG nova.compute.manager [req-5b9803bb-2a46-4a74-bc44-d82282703f9f req-672af96c-d32f-422f-a93f-1e75e0a77d55 service nova] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Refreshing instance network info cache due to event network-changed-3dae35ba-bed5-44cb-a046-47bcc1cdc551. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 685.666242] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b9803bb-2a46-4a74-bc44-d82282703f9f req-672af96c-d32f-422f-a93f-1e75e0a77d55 service nova] Acquiring lock "refresh_cache-2f5bd366-317e-4c18-956a-35d9cdbae9e2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.666242] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b9803bb-2a46-4a74-bc44-d82282703f9f req-672af96c-d32f-422f-a93f-1e75e0a77d55 service nova] Acquired lock "refresh_cache-2f5bd366-317e-4c18-956a-35d9cdbae9e2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.666242] env[61906]: DEBUG nova.network.neutron [req-5b9803bb-2a46-4a74-bc44-d82282703f9f req-672af96c-d32f-422f-a93f-1e75e0a77d55 service nova] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Refreshing network info cache for port 3dae35ba-bed5-44cb-a046-47bcc1cdc551 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 685.805717] env[61906]: DEBUG nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 685.865054] env[61906]: ERROR nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3dae35ba-bed5-44cb-a046-47bcc1cdc551, please check neutron logs for more information. [ 685.865054] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 685.865054] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.865054] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 685.865054] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.865054] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 685.865054] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.865054] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 685.865054] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.865054] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 685.865054] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.865054] env[61906]: ERROR nova.compute.manager raise self.value [ 685.865054] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.865054] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 685.865054] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.865054] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 685.865744] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.865744] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 685.865744] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3dae35ba-bed5-44cb-a046-47bcc1cdc551, please check neutron logs for more information. [ 685.865744] env[61906]: ERROR nova.compute.manager [ 685.865744] env[61906]: Traceback (most recent call last): [ 685.865744] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 685.865744] env[61906]: listener.cb(fileno) [ 685.865744] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.865744] env[61906]: result = function(*args, **kwargs) [ 685.865744] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 685.865744] env[61906]: return func(*args, **kwargs) [ 685.865744] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.865744] env[61906]: raise e [ 685.865744] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.865744] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 685.865744] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.865744] env[61906]: created_port_ids = self._update_ports_for_instance( [ 685.865744] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.865744] env[61906]: with excutils.save_and_reraise_exception(): [ 685.865744] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.865744] env[61906]: self.force_reraise() [ 685.865744] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.865744] env[61906]: raise self.value [ 685.865744] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.865744] env[61906]: updated_port = self._update_port( [ 685.865744] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.865744] env[61906]: _ensure_no_port_binding_failure(port) [ 685.865744] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.865744] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 685.866844] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 3dae35ba-bed5-44cb-a046-47bcc1cdc551, please check neutron logs for more information. [ 685.866844] env[61906]: Removing descriptor: 15 [ 685.913722] env[61906]: DEBUG nova.scheduler.client.report [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.984792] env[61906]: DEBUG nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 686.013391] env[61906]: DEBUG nova.virt.hardware [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 686.013665] env[61906]: DEBUG nova.virt.hardware [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 686.013821] env[61906]: DEBUG nova.virt.hardware [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 686.014009] env[61906]: DEBUG nova.virt.hardware [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 686.014167] env[61906]: DEBUG nova.virt.hardware [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 686.014314] env[61906]: DEBUG nova.virt.hardware [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 686.014525] env[61906]: DEBUG nova.virt.hardware [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 686.014673] env[61906]: DEBUG nova.virt.hardware [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 686.014867] env[61906]: DEBUG nova.virt.hardware [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 686.015107] env[61906]: DEBUG nova.virt.hardware [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 686.015327] env[61906]: DEBUG nova.virt.hardware [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.016267] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b88daf5-6057-4aff-9ea9-676e5c7655c6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.024233] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b7e8a7-ffc6-4f04-991e-a464d9c28dde {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.037894] env[61906]: ERROR nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3dae35ba-bed5-44cb-a046-47bcc1cdc551, please check neutron logs for more information. [ 686.037894] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Traceback (most recent call last): [ 686.037894] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 686.037894] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] yield resources [ 686.037894] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 686.037894] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] self.driver.spawn(context, instance, image_meta, [ 686.037894] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 686.037894] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.037894] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.037894] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] vm_ref = self.build_virtual_machine(instance, [ 686.037894] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.038272] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.038272] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.038272] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] for vif in network_info: [ 686.038272] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 686.038272] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] return self._sync_wrapper(fn, *args, **kwargs) [ 686.038272] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 686.038272] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] self.wait() [ 686.038272] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 686.038272] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] self[:] = self._gt.wait() [ 686.038272] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.038272] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] return self._exit_event.wait() [ 686.038272] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 686.038272] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] current.throw(*self._exc) [ 686.038593] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.038593] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] result = function(*args, **kwargs) [ 686.038593] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 686.038593] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] return func(*args, **kwargs) [ 686.038593] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.038593] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] raise e [ 686.038593] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.038593] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] nwinfo = self.network_api.allocate_for_instance( [ 686.038593] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.038593] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] created_port_ids = self._update_ports_for_instance( [ 686.038593] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.038593] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] with excutils.save_and_reraise_exception(): [ 686.038593] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.038971] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] self.force_reraise() [ 686.038971] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.038971] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] raise self.value [ 686.038971] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.038971] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] updated_port = self._update_port( [ 686.038971] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.038971] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] _ensure_no_port_binding_failure(port) [ 686.038971] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.038971] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] raise exception.PortBindingFailed(port_id=port['id']) [ 686.038971] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] nova.exception.PortBindingFailed: Binding failed for port 3dae35ba-bed5-44cb-a046-47bcc1cdc551, please check neutron logs for more information. [ 686.038971] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] [ 686.038971] env[61906]: INFO nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Terminating instance [ 686.040456] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Acquiring lock "refresh_cache-2f5bd366-317e-4c18-956a-35d9cdbae9e2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.186153] env[61906]: DEBUG nova.network.neutron [req-5b9803bb-2a46-4a74-bc44-d82282703f9f req-672af96c-d32f-422f-a93f-1e75e0a77d55 service nova] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.327402] env[61906]: DEBUG nova.network.neutron [req-5b9803bb-2a46-4a74-bc44-d82282703f9f req-672af96c-d32f-422f-a93f-1e75e0a77d55 service nova] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.348069] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.419253] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.419686] env[61906]: DEBUG nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 686.422774] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.136s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.832180] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b9803bb-2a46-4a74-bc44-d82282703f9f req-672af96c-d32f-422f-a93f-1e75e0a77d55 service nova] Releasing lock "refresh_cache-2f5bd366-317e-4c18-956a-35d9cdbae9e2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.832604] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Acquired lock "refresh_cache-2f5bd366-317e-4c18-956a-35d9cdbae9e2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.832790] env[61906]: DEBUG nova.network.neutron [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 686.927625] env[61906]: DEBUG nova.compute.utils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 686.932333] env[61906]: DEBUG nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 686.932514] env[61906]: DEBUG nova.network.neutron [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 686.995034] env[61906]: DEBUG nova.policy [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2d5cabf716441a2aef356b324035b73', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '717c30385610400793faa971da525812', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 687.260966] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03088797-73bc-40b2-b971-b93b6df20f61 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.268269] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2019c79e-8c42-4620-8caa-8b738b7577bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.298816] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2cd7ee-1fde-4f60-b1e1-0e105ac1fc97 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.306653] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a80515-c6de-43e8-a305-a91505df7c99 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.320628] env[61906]: DEBUG nova.compute.provider_tree [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.358182] env[61906]: DEBUG nova.network.neutron [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.406537] env[61906]: DEBUG nova.network.neutron [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Successfully created port: e023640e-63ff-46e3-bc0c-7ccc3d59c920 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 687.433321] env[61906]: DEBUG nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 687.444513] env[61906]: DEBUG nova.network.neutron [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.722676] env[61906]: DEBUG nova.compute.manager [req-22ea2c8e-4a7b-4bcf-bde9-4b58739b439c req-bfe396f3-47af-4cfe-a7dd-95b7e9c79ca3 service nova] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Received event network-vif-deleted-3dae35ba-bed5-44cb-a046-47bcc1cdc551 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 687.786725] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquiring lock "c3cb1a2d-e905-497c-a4de-8c6579c37156" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.786998] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Lock "c3cb1a2d-e905-497c-a4de-8c6579c37156" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.823426] env[61906]: DEBUG nova.scheduler.client.report [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.949650] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Releasing lock "refresh_cache-2f5bd366-317e-4c18-956a-35d9cdbae9e2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.949895] env[61906]: DEBUG nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 687.950012] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 687.950314] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c752189-8da8-4226-bd43-c71649499946 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.959682] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8c6a2a-f5bd-47ad-ab75-b26d71cc94e1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.982199] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2f5bd366-317e-4c18-956a-35d9cdbae9e2 could not be found. [ 687.982414] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 687.982586] env[61906]: INFO nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 687.982826] env[61906]: DEBUG oslo.service.loopingcall [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.983045] env[61906]: DEBUG nova.compute.manager [-] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.983134] env[61906]: DEBUG nova.network.neutron [-] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 687.997070] env[61906]: DEBUG nova.network.neutron [-] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.329601] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.907s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.330242] env[61906]: ERROR nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f1de37fc-eff3-425a-a263-b13b723e3323, please check neutron logs for more information. [ 688.330242] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Traceback (most recent call last): [ 688.330242] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 688.330242] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] self.driver.spawn(context, instance, image_meta, [ 688.330242] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 688.330242] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 688.330242] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 688.330242] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] vm_ref = self.build_virtual_machine(instance, [ 688.330242] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 688.330242] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] vif_infos = vmwarevif.get_vif_info(self._session, [ 688.330242] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 688.330549] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] for vif in network_info: [ 688.330549] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 688.330549] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] return self._sync_wrapper(fn, *args, **kwargs) [ 688.330549] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 688.330549] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] self.wait() [ 688.330549] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 688.330549] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] self[:] = self._gt.wait() [ 688.330549] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 688.330549] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] return self._exit_event.wait() [ 688.330549] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 688.330549] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] result = hub.switch() [ 688.330549] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 688.330549] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] return self.greenlet.switch() [ 688.330891] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.330891] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] result = function(*args, **kwargs) [ 688.330891] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 688.330891] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] return func(*args, **kwargs) [ 688.330891] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.330891] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] raise e [ 688.330891] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.330891] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] nwinfo = self.network_api.allocate_for_instance( [ 688.330891] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.330891] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] created_port_ids = self._update_ports_for_instance( [ 688.330891] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.330891] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] with excutils.save_and_reraise_exception(): [ 688.330891] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.331229] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] self.force_reraise() [ 688.331229] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.331229] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] raise self.value [ 688.331229] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.331229] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] updated_port = self._update_port( [ 688.331229] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.331229] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] _ensure_no_port_binding_failure(port) [ 688.331229] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.331229] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] raise exception.PortBindingFailed(port_id=port['id']) [ 688.331229] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] nova.exception.PortBindingFailed: Binding failed for port f1de37fc-eff3-425a-a263-b13b723e3323, please check neutron logs for more information. [ 688.331229] env[61906]: ERROR nova.compute.manager [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] [ 688.331507] env[61906]: DEBUG nova.compute.utils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Binding failed for port f1de37fc-eff3-425a-a263-b13b723e3323, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 688.332310] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.570s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.333801] env[61906]: INFO nova.compute.claims [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 688.336315] env[61906]: DEBUG nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Build of instance 95604a75-0ab3-45d9-a235-9067e615e72a was re-scheduled: Binding failed for port f1de37fc-eff3-425a-a263-b13b723e3323, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 688.337020] env[61906]: DEBUG nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 688.337020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Acquiring lock "refresh_cache-95604a75-0ab3-45d9-a235-9067e615e72a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.337128] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Acquired lock "refresh_cache-95604a75-0ab3-45d9-a235-9067e615e72a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.337290] env[61906]: DEBUG nova.network.neutron [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.410197] env[61906]: ERROR nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e023640e-63ff-46e3-bc0c-7ccc3d59c920, please check neutron logs for more information. [ 688.410197] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 688.410197] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.410197] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 688.410197] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.410197] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 688.410197] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.410197] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 688.410197] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.410197] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 688.410197] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.410197] env[61906]: ERROR nova.compute.manager raise self.value [ 688.410197] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.410197] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 688.410197] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.410197] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 688.410781] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.410781] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 688.410781] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e023640e-63ff-46e3-bc0c-7ccc3d59c920, please check neutron logs for more information. [ 688.410781] env[61906]: ERROR nova.compute.manager [ 688.410781] env[61906]: Traceback (most recent call last): [ 688.410781] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 688.410781] env[61906]: listener.cb(fileno) [ 688.410781] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.410781] env[61906]: result = function(*args, **kwargs) [ 688.410781] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 688.410781] env[61906]: return func(*args, **kwargs) [ 688.410781] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.410781] env[61906]: raise e [ 688.410781] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.410781] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 688.410781] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.410781] env[61906]: created_port_ids = self._update_ports_for_instance( [ 688.410781] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.410781] env[61906]: with excutils.save_and_reraise_exception(): [ 688.410781] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.410781] env[61906]: self.force_reraise() [ 688.410781] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.410781] env[61906]: raise self.value [ 688.410781] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.410781] env[61906]: updated_port = self._update_port( [ 688.410781] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.410781] env[61906]: _ensure_no_port_binding_failure(port) [ 688.410781] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.410781] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 688.411665] env[61906]: nova.exception.PortBindingFailed: Binding failed for port e023640e-63ff-46e3-bc0c-7ccc3d59c920, please check neutron logs for more information. [ 688.411665] env[61906]: Removing descriptor: 15 [ 688.442254] env[61906]: DEBUG nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 688.469012] env[61906]: DEBUG nova.virt.hardware [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 688.469012] env[61906]: DEBUG nova.virt.hardware [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 688.469012] env[61906]: DEBUG nova.virt.hardware [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 688.469012] env[61906]: DEBUG nova.virt.hardware [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 688.469213] env[61906]: DEBUG nova.virt.hardware [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 688.469213] env[61906]: DEBUG nova.virt.hardware [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 688.469213] env[61906]: DEBUG nova.virt.hardware [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 688.469213] env[61906]: DEBUG nova.virt.hardware [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 688.469213] env[61906]: DEBUG nova.virt.hardware [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 688.469382] env[61906]: DEBUG nova.virt.hardware [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 688.469382] env[61906]: DEBUG nova.virt.hardware [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.469794] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b5790a-993a-4af3-a464-74645bf6f9ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.479353] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbc82b2-3bef-464a-86b3-7c3d29d4bff6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.494851] env[61906]: ERROR nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e023640e-63ff-46e3-bc0c-7ccc3d59c920, please check neutron logs for more information. [ 688.494851] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Traceback (most recent call last): [ 688.494851] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 688.494851] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] yield resources [ 688.494851] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 688.494851] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] self.driver.spawn(context, instance, image_meta, [ 688.494851] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 688.494851] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 688.494851] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 688.494851] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] vm_ref = self.build_virtual_machine(instance, [ 688.494851] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 688.495316] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] vif_infos = vmwarevif.get_vif_info(self._session, [ 688.495316] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 688.495316] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] for vif in network_info: [ 688.495316] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 688.495316] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] return self._sync_wrapper(fn, *args, **kwargs) [ 688.495316] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 688.495316] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] self.wait() [ 688.495316] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 688.495316] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] self[:] = self._gt.wait() [ 688.495316] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 688.495316] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] return self._exit_event.wait() [ 688.495316] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 688.495316] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] current.throw(*self._exc) [ 688.495727] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.495727] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] result = function(*args, **kwargs) [ 688.495727] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 688.495727] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] return func(*args, **kwargs) [ 688.495727] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.495727] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] raise e [ 688.495727] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.495727] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] nwinfo = self.network_api.allocate_for_instance( [ 688.495727] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.495727] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] created_port_ids = self._update_ports_for_instance( [ 688.495727] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.495727] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] with excutils.save_and_reraise_exception(): [ 688.495727] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.496130] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] self.force_reraise() [ 688.496130] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.496130] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] raise self.value [ 688.496130] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.496130] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] updated_port = self._update_port( [ 688.496130] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.496130] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] _ensure_no_port_binding_failure(port) [ 688.496130] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.496130] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] raise exception.PortBindingFailed(port_id=port['id']) [ 688.496130] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] nova.exception.PortBindingFailed: Binding failed for port e023640e-63ff-46e3-bc0c-7ccc3d59c920, please check neutron logs for more information. [ 688.496130] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] [ 688.496130] env[61906]: INFO nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Terminating instance [ 688.497344] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Acquiring lock "refresh_cache-cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.497504] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Acquired lock "refresh_cache-cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.497675] env[61906]: DEBUG nova.network.neutron [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.500138] env[61906]: DEBUG nova.network.neutron [-] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.855988] env[61906]: DEBUG nova.network.neutron [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.952534] env[61906]: DEBUG nova.network.neutron [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.002272] env[61906]: INFO nova.compute.manager [-] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Took 1.02 seconds to deallocate network for instance. [ 689.005243] env[61906]: DEBUG nova.compute.claims [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 689.005419] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.017955] env[61906]: DEBUG nova.network.neutron [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.073672] env[61906]: DEBUG nova.network.neutron [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.454913] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Releasing lock "refresh_cache-95604a75-0ab3-45d9-a235-9067e615e72a" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.455162] env[61906]: DEBUG nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 689.455366] env[61906]: DEBUG nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.455554] env[61906]: DEBUG nova.network.neutron [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.487551] env[61906]: DEBUG nova.network.neutron [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.577469] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Releasing lock "refresh_cache-cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.577908] env[61906]: DEBUG nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 689.578151] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 689.578605] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e593a633-41d1-47f3-9309-f68104375aad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.587691] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628f8af8-9eed-4ee9-8cec-4762c66614ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.613047] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b could not be found. [ 689.613277] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 689.613516] env[61906]: INFO nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 689.613774] env[61906]: DEBUG oslo.service.loopingcall [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.616703] env[61906]: DEBUG nova.compute.manager [-] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.616807] env[61906]: DEBUG nova.network.neutron [-] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.632977] env[61906]: DEBUG nova.network.neutron [-] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.662353] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f404b6-54bc-4288-b074-a67f533967ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.670200] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a98963-06d5-4620-aca1-13c69f412b56 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.700689] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6f5471-fc18-436a-b6de-15918f3cbe00 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.708184] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98ac3e5-ae10-4129-8d59-15cc0485903b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.722174] env[61906]: DEBUG nova.compute.provider_tree [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.746569] env[61906]: DEBUG nova.compute.manager [req-368f97b1-896f-4760-8393-5fa2e41b1e5b req-6ae04936-62b7-4a58-b4c6-ad58fe121e9d service nova] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Received event network-changed-e023640e-63ff-46e3-bc0c-7ccc3d59c920 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 689.746778] env[61906]: DEBUG nova.compute.manager [req-368f97b1-896f-4760-8393-5fa2e41b1e5b req-6ae04936-62b7-4a58-b4c6-ad58fe121e9d service nova] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Refreshing instance network info cache due to event network-changed-e023640e-63ff-46e3-bc0c-7ccc3d59c920. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 689.746990] env[61906]: DEBUG oslo_concurrency.lockutils [req-368f97b1-896f-4760-8393-5fa2e41b1e5b req-6ae04936-62b7-4a58-b4c6-ad58fe121e9d service nova] Acquiring lock "refresh_cache-cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.747165] env[61906]: DEBUG oslo_concurrency.lockutils [req-368f97b1-896f-4760-8393-5fa2e41b1e5b req-6ae04936-62b7-4a58-b4c6-ad58fe121e9d service nova] Acquired lock "refresh_cache-cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.747301] env[61906]: DEBUG nova.network.neutron [req-368f97b1-896f-4760-8393-5fa2e41b1e5b req-6ae04936-62b7-4a58-b4c6-ad58fe121e9d service nova] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Refreshing network info cache for port e023640e-63ff-46e3-bc0c-7ccc3d59c920 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 689.992387] env[61906]: DEBUG nova.network.neutron [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.135277] env[61906]: DEBUG nova.network.neutron [-] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.225044] env[61906]: DEBUG nova.scheduler.client.report [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.264288] env[61906]: DEBUG nova.network.neutron [req-368f97b1-896f-4760-8393-5fa2e41b1e5b req-6ae04936-62b7-4a58-b4c6-ad58fe121e9d service nova] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.348046] env[61906]: DEBUG nova.network.neutron [req-368f97b1-896f-4760-8393-5fa2e41b1e5b req-6ae04936-62b7-4a58-b4c6-ad58fe121e9d service nova] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.494918] env[61906]: INFO nova.compute.manager [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] [instance: 95604a75-0ab3-45d9-a235-9067e615e72a] Took 1.04 seconds to deallocate network for instance. [ 690.637638] env[61906]: INFO nova.compute.manager [-] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Took 1.02 seconds to deallocate network for instance. [ 690.639895] env[61906]: DEBUG nova.compute.claims [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 690.640095] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.729607] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.730154] env[61906]: DEBUG nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 690.733264] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.486s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.734737] env[61906]: INFO nova.compute.claims [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.853879] env[61906]: DEBUG oslo_concurrency.lockutils [req-368f97b1-896f-4760-8393-5fa2e41b1e5b req-6ae04936-62b7-4a58-b4c6-ad58fe121e9d service nova] Releasing lock "refresh_cache-cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.853879] env[61906]: DEBUG nova.compute.manager [req-368f97b1-896f-4760-8393-5fa2e41b1e5b req-6ae04936-62b7-4a58-b4c6-ad58fe121e9d service nova] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Received event network-vif-deleted-e023640e-63ff-46e3-bc0c-7ccc3d59c920 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 691.237021] env[61906]: DEBUG nova.compute.utils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 691.238231] env[61906]: DEBUG nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 691.238565] env[61906]: DEBUG nova.network.neutron [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 691.279826] env[61906]: DEBUG nova.policy [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'abd858787f6945349a08939438d73518', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fde398d0f99d4033b15e86d479855c50', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 691.526812] env[61906]: INFO nova.scheduler.client.report [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Deleted allocations for instance 95604a75-0ab3-45d9-a235-9067e615e72a [ 691.542082] env[61906]: DEBUG nova.network.neutron [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Successfully created port: 3a5c8162-640f-461e-9419-6593c9dad395 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 691.744269] env[61906]: DEBUG nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 692.035663] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1091da0-d826-4f73-8d74-e9f99368df79 tempest-ServersTestManualDisk-596290628 tempest-ServersTestManualDisk-596290628-project-member] Lock "95604a75-0ab3-45d9-a235-9067e615e72a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.937s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.036510] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d39bd55-ad0a-440e-a66e-1438f195b452 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.045289] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8c01d3-9fdc-466d-b99a-4aac11cc0332 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.076294] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be16686c-dae4-47c3-bc54-c87710814407 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.084764] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956af424-7c2c-4194-91e4-6721482d46eb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.098838] env[61906]: DEBUG nova.compute.provider_tree [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.298481] env[61906]: DEBUG nova.compute.manager [req-cd342ebe-3383-47c7-bbf1-a7e2af083c1b req-a5d71236-8853-4a12-b31a-0e5f8c253ff2 service nova] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Received event network-changed-3a5c8162-640f-461e-9419-6593c9dad395 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 692.298732] env[61906]: DEBUG nova.compute.manager [req-cd342ebe-3383-47c7-bbf1-a7e2af083c1b req-a5d71236-8853-4a12-b31a-0e5f8c253ff2 service nova] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Refreshing instance network info cache due to event network-changed-3a5c8162-640f-461e-9419-6593c9dad395. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 692.298887] env[61906]: DEBUG oslo_concurrency.lockutils [req-cd342ebe-3383-47c7-bbf1-a7e2af083c1b req-a5d71236-8853-4a12-b31a-0e5f8c253ff2 service nova] Acquiring lock "refresh_cache-746ec539-5e13-4a9f-af73-26b175ad1e41" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.299039] env[61906]: DEBUG oslo_concurrency.lockutils [req-cd342ebe-3383-47c7-bbf1-a7e2af083c1b req-a5d71236-8853-4a12-b31a-0e5f8c253ff2 service nova] Acquired lock "refresh_cache-746ec539-5e13-4a9f-af73-26b175ad1e41" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.299196] env[61906]: DEBUG nova.network.neutron [req-cd342ebe-3383-47c7-bbf1-a7e2af083c1b req-a5d71236-8853-4a12-b31a-0e5f8c253ff2 service nova] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Refreshing network info cache for port 3a5c8162-640f-461e-9419-6593c9dad395 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 692.495842] env[61906]: ERROR nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3a5c8162-640f-461e-9419-6593c9dad395, please check neutron logs for more information. [ 692.495842] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 692.495842] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.495842] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 692.495842] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.495842] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 692.495842] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.495842] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 692.495842] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.495842] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 692.495842] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.495842] env[61906]: ERROR nova.compute.manager raise self.value [ 692.495842] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.495842] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 692.495842] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.495842] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 692.496424] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.496424] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 692.496424] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3a5c8162-640f-461e-9419-6593c9dad395, please check neutron logs for more information. [ 692.496424] env[61906]: ERROR nova.compute.manager [ 692.496424] env[61906]: Traceback (most recent call last): [ 692.496424] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 692.496424] env[61906]: listener.cb(fileno) [ 692.496424] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.496424] env[61906]: result = function(*args, **kwargs) [ 692.496424] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 692.496424] env[61906]: return func(*args, **kwargs) [ 692.496424] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.496424] env[61906]: raise e [ 692.496424] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.496424] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 692.496424] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.496424] env[61906]: created_port_ids = self._update_ports_for_instance( [ 692.496424] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.496424] env[61906]: with excutils.save_and_reraise_exception(): [ 692.496424] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.496424] env[61906]: self.force_reraise() [ 692.496424] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.496424] env[61906]: raise self.value [ 692.496424] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.496424] env[61906]: updated_port = self._update_port( [ 692.496424] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.496424] env[61906]: _ensure_no_port_binding_failure(port) [ 692.496424] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.496424] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 692.497293] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 3a5c8162-640f-461e-9419-6593c9dad395, please check neutron logs for more information. [ 692.497293] env[61906]: Removing descriptor: 15 [ 692.542038] env[61906]: DEBUG nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 692.605481] env[61906]: DEBUG nova.scheduler.client.report [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.760624] env[61906]: DEBUG nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 692.786280] env[61906]: DEBUG nova.virt.hardware [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 692.786549] env[61906]: DEBUG nova.virt.hardware [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 692.786710] env[61906]: DEBUG nova.virt.hardware [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.786887] env[61906]: DEBUG nova.virt.hardware [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 692.787062] env[61906]: DEBUG nova.virt.hardware [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.787667] env[61906]: DEBUG nova.virt.hardware [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 692.787667] env[61906]: DEBUG nova.virt.hardware [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 692.787667] env[61906]: DEBUG nova.virt.hardware [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 692.787817] env[61906]: DEBUG nova.virt.hardware [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 692.787886] env[61906]: DEBUG nova.virt.hardware [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 692.788073] env[61906]: DEBUG nova.virt.hardware [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.788927] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc157ea-7cd2-4549-bc57-81edddc51412 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.797990] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf48f48e-caec-4fa1-bf90-ca299c31e4ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.813614] env[61906]: ERROR nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3a5c8162-640f-461e-9419-6593c9dad395, please check neutron logs for more information. [ 692.813614] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Traceback (most recent call last): [ 692.813614] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 692.813614] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] yield resources [ 692.813614] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 692.813614] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] self.driver.spawn(context, instance, image_meta, [ 692.813614] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 692.813614] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.813614] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.813614] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] vm_ref = self.build_virtual_machine(instance, [ 692.813614] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.813968] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.813968] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.813968] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] for vif in network_info: [ 692.813968] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.813968] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] return self._sync_wrapper(fn, *args, **kwargs) [ 692.813968] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.813968] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] self.wait() [ 692.813968] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.813968] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] self[:] = self._gt.wait() [ 692.813968] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.813968] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] return self._exit_event.wait() [ 692.813968] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 692.813968] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] current.throw(*self._exc) [ 692.814358] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.814358] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] result = function(*args, **kwargs) [ 692.814358] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 692.814358] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] return func(*args, **kwargs) [ 692.814358] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.814358] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] raise e [ 692.814358] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.814358] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] nwinfo = self.network_api.allocate_for_instance( [ 692.814358] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.814358] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] created_port_ids = self._update_ports_for_instance( [ 692.814358] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.814358] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] with excutils.save_and_reraise_exception(): [ 692.814358] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.815174] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] self.force_reraise() [ 692.815174] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.815174] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] raise self.value [ 692.815174] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.815174] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] updated_port = self._update_port( [ 692.815174] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.815174] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] _ensure_no_port_binding_failure(port) [ 692.815174] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.815174] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] raise exception.PortBindingFailed(port_id=port['id']) [ 692.815174] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] nova.exception.PortBindingFailed: Binding failed for port 3a5c8162-640f-461e-9419-6593c9dad395, please check neutron logs for more information. [ 692.815174] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] [ 692.815174] env[61906]: INFO nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Terminating instance [ 692.815586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquiring lock "refresh_cache-746ec539-5e13-4a9f-af73-26b175ad1e41" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.830848] env[61906]: DEBUG nova.network.neutron [req-cd342ebe-3383-47c7-bbf1-a7e2af083c1b req-a5d71236-8853-4a12-b31a-0e5f8c253ff2 service nova] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.982147] env[61906]: DEBUG nova.network.neutron [req-cd342ebe-3383-47c7-bbf1-a7e2af083c1b req-a5d71236-8853-4a12-b31a-0e5f8c253ff2 service nova] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.064750] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.112463] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.113370] env[61906]: DEBUG nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 693.119448] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.169s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.484428] env[61906]: DEBUG oslo_concurrency.lockutils [req-cd342ebe-3383-47c7-bbf1-a7e2af083c1b req-a5d71236-8853-4a12-b31a-0e5f8c253ff2 service nova] Releasing lock "refresh_cache-746ec539-5e13-4a9f-af73-26b175ad1e41" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.484428] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquired lock "refresh_cache-746ec539-5e13-4a9f-af73-26b175ad1e41" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.484428] env[61906]: DEBUG nova.network.neutron [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.630136] env[61906]: DEBUG nova.compute.utils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.632457] env[61906]: DEBUG nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.632457] env[61906]: DEBUG nova.network.neutron [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 693.701195] env[61906]: DEBUG nova.policy [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf4fea47b929461785f4a487a56977a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3093466de87a45a2ac673c25bea19bc3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.971875] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bbb5581-cd48-4a3a-b3a9-9a40157cbcb3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.979172] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7a29fe-3fbc-4cfc-be05-aea623868d1e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.023206] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eab54d50-b914-451b-ac5b-9e6248670940 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.034887] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f84b2779-eae9-40a3-a06c-005efb24888b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.049774] env[61906]: DEBUG nova.compute.provider_tree [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.056162] env[61906]: DEBUG nova.network.neutron [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.136081] env[61906]: DEBUG nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 694.180746] env[61906]: DEBUG nova.network.neutron [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.189140] env[61906]: DEBUG nova.network.neutron [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Successfully created port: e107c8dc-cca7-47c3-b7ec-5d7c24a24f36 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 694.481325] env[61906]: DEBUG nova.compute.manager [req-cd2eaa86-b983-4498-a1a8-fb9eda394338 req-ea325700-22ac-474d-a854-b2147492f4b7 service nova] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Received event network-vif-deleted-3a5c8162-640f-461e-9419-6593c9dad395 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.552790] env[61906]: DEBUG nova.scheduler.client.report [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.683987] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Releasing lock "refresh_cache-746ec539-5e13-4a9f-af73-26b175ad1e41" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.684412] env[61906]: DEBUG nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 694.684819] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 694.685202] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d03db9bf-9805-4016-b91a-9d5e0c49204d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.695593] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44025f57-47d4-47b2-9a4c-d328c3851ee4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.717437] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 746ec539-5e13-4a9f-af73-26b175ad1e41 could not be found. [ 694.717643] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 694.717822] env[61906]: INFO nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Took 0.03 seconds to destroy the instance on the hypervisor. [ 694.718072] env[61906]: DEBUG oslo.service.loopingcall [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 694.718282] env[61906]: DEBUG nova.compute.manager [-] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.718373] env[61906]: DEBUG nova.network.neutron [-] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 694.732909] env[61906]: DEBUG nova.network.neutron [-] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.057838] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.938s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.058520] env[61906]: ERROR nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 82b16bf3-088a-4900-a616-c820e33458dd, please check neutron logs for more information. [ 695.058520] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Traceback (most recent call last): [ 695.058520] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 695.058520] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] self.driver.spawn(context, instance, image_meta, [ 695.058520] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 695.058520] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.058520] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.058520] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] vm_ref = self.build_virtual_machine(instance, [ 695.058520] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.058520] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.058520] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] for vif in network_info: [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] return self._sync_wrapper(fn, *args, **kwargs) [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] self.wait() [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] self[:] = self._gt.wait() [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] return self._exit_event.wait() [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] current.throw(*self._exc) [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] result = function(*args, **kwargs) [ 695.058966] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 695.059489] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] return func(*args, **kwargs) [ 695.059489] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.059489] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] raise e [ 695.059489] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.059489] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] nwinfo = self.network_api.allocate_for_instance( [ 695.059489] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.059489] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] created_port_ids = self._update_ports_for_instance( [ 695.059489] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.059489] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] with excutils.save_and_reraise_exception(): [ 695.059489] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.059489] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] self.force_reraise() [ 695.059489] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.059489] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] raise self.value [ 695.060071] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.060071] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] updated_port = self._update_port( [ 695.060071] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.060071] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] _ensure_no_port_binding_failure(port) [ 695.060071] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.060071] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] raise exception.PortBindingFailed(port_id=port['id']) [ 695.060071] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] nova.exception.PortBindingFailed: Binding failed for port 82b16bf3-088a-4900-a616-c820e33458dd, please check neutron logs for more information. [ 695.060071] env[61906]: ERROR nova.compute.manager [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] [ 695.060071] env[61906]: DEBUG nova.compute.utils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Binding failed for port 82b16bf3-088a-4900-a616-c820e33458dd, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 695.060985] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.484s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.064449] env[61906]: DEBUG nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Build of instance 42135fac-aef7-47d9-9afe-5ac2351bb07c was re-scheduled: Binding failed for port 82b16bf3-088a-4900-a616-c820e33458dd, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 695.065020] env[61906]: DEBUG nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 695.065253] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "refresh_cache-42135fac-aef7-47d9-9afe-5ac2351bb07c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.065399] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquired lock "refresh_cache-42135fac-aef7-47d9-9afe-5ac2351bb07c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.065559] env[61906]: DEBUG nova.network.neutron [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.144710] env[61906]: DEBUG nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 695.179313] env[61906]: DEBUG nova.virt.hardware [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 695.179799] env[61906]: DEBUG nova.virt.hardware [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 695.179799] env[61906]: DEBUG nova.virt.hardware [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.179975] env[61906]: DEBUG nova.virt.hardware [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 695.180426] env[61906]: DEBUG nova.virt.hardware [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.180426] env[61906]: DEBUG nova.virt.hardware [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 695.180547] env[61906]: DEBUG nova.virt.hardware [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 695.180660] env[61906]: DEBUG nova.virt.hardware [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 695.180811] env[61906]: DEBUG nova.virt.hardware [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 695.180975] env[61906]: DEBUG nova.virt.hardware [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 695.181166] env[61906]: DEBUG nova.virt.hardware [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 695.182200] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a019726-966a-4b72-8e68-df0edb1d8a2f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.194481] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698af207-d183-4072-b044-3967eadd6cca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.235714] env[61906]: DEBUG nova.network.neutron [-] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.577521] env[61906]: ERROR nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e107c8dc-cca7-47c3-b7ec-5d7c24a24f36, please check neutron logs for more information. [ 695.577521] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 695.577521] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.577521] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 695.577521] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.577521] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 695.577521] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.577521] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 695.577521] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.577521] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 695.577521] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.577521] env[61906]: ERROR nova.compute.manager raise self.value [ 695.577521] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.577521] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 695.577521] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.577521] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 695.578242] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.578242] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 695.578242] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e107c8dc-cca7-47c3-b7ec-5d7c24a24f36, please check neutron logs for more information. [ 695.578242] env[61906]: ERROR nova.compute.manager [ 695.578242] env[61906]: Traceback (most recent call last): [ 695.578242] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 695.578242] env[61906]: listener.cb(fileno) [ 695.578242] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.578242] env[61906]: result = function(*args, **kwargs) [ 695.578242] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 695.578242] env[61906]: return func(*args, **kwargs) [ 695.578242] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.578242] env[61906]: raise e [ 695.578242] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.578242] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 695.578242] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.578242] env[61906]: created_port_ids = self._update_ports_for_instance( [ 695.578242] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.578242] env[61906]: with excutils.save_and_reraise_exception(): [ 695.578242] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.578242] env[61906]: self.force_reraise() [ 695.578242] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.578242] env[61906]: raise self.value [ 695.578242] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.578242] env[61906]: updated_port = self._update_port( [ 695.578242] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.578242] env[61906]: _ensure_no_port_binding_failure(port) [ 695.578242] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.578242] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 695.579094] env[61906]: nova.exception.PortBindingFailed: Binding failed for port e107c8dc-cca7-47c3-b7ec-5d7c24a24f36, please check neutron logs for more information. [ 695.579094] env[61906]: Removing descriptor: 15 [ 695.579094] env[61906]: ERROR nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e107c8dc-cca7-47c3-b7ec-5d7c24a24f36, please check neutron logs for more information. [ 695.579094] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Traceback (most recent call last): [ 695.579094] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 695.579094] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] yield resources [ 695.579094] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 695.579094] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] self.driver.spawn(context, instance, image_meta, [ 695.579094] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 695.579094] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.579094] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.579094] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] vm_ref = self.build_virtual_machine(instance, [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] for vif in network_info: [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] return self._sync_wrapper(fn, *args, **kwargs) [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] self.wait() [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] self[:] = self._gt.wait() [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] return self._exit_event.wait() [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 695.579447] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] result = hub.switch() [ 695.579840] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 695.579840] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] return self.greenlet.switch() [ 695.579840] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.579840] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] result = function(*args, **kwargs) [ 695.579840] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 695.579840] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] return func(*args, **kwargs) [ 695.579840] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.579840] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] raise e [ 695.579840] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.579840] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] nwinfo = self.network_api.allocate_for_instance( [ 695.579840] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.579840] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] created_port_ids = self._update_ports_for_instance( [ 695.579840] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.580215] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] with excutils.save_and_reraise_exception(): [ 695.580215] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.580215] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] self.force_reraise() [ 695.580215] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.580215] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] raise self.value [ 695.580215] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.580215] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] updated_port = self._update_port( [ 695.580215] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.580215] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] _ensure_no_port_binding_failure(port) [ 695.580215] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.580215] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] raise exception.PortBindingFailed(port_id=port['id']) [ 695.580215] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] nova.exception.PortBindingFailed: Binding failed for port e107c8dc-cca7-47c3-b7ec-5d7c24a24f36, please check neutron logs for more information. [ 695.580215] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] [ 695.580570] env[61906]: INFO nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Terminating instance [ 695.580826] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "refresh_cache-40dae609-e7a5-4af2-84d9-a53f93d95743" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.580993] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "refresh_cache-40dae609-e7a5-4af2-84d9-a53f93d95743" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.581173] env[61906]: DEBUG nova.network.neutron [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.591268] env[61906]: DEBUG nova.network.neutron [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.673931] env[61906]: DEBUG nova.network.neutron [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.740195] env[61906]: INFO nova.compute.manager [-] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Took 1.02 seconds to deallocate network for instance. [ 695.744965] env[61906]: DEBUG nova.compute.claims [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 695.745164] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.880477] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db3bac45-61ff-4bf2-aaaf-9cb7c09575ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.888713] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26cb797d-fb6f-4251-bd2b-a1a6e701b18c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.918585] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2df393-3cfa-46d7-82c0-8c86fe435d3b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.926846] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d36161-29b8-4934-a955-6982c0fc3f8b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.939369] env[61906]: DEBUG nova.compute.provider_tree [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.101545] env[61906]: DEBUG nova.network.neutron [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.177952] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Releasing lock "refresh_cache-42135fac-aef7-47d9-9afe-5ac2351bb07c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.177952] env[61906]: DEBUG nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 696.177952] env[61906]: DEBUG nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.177952] env[61906]: DEBUG nova.network.neutron [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.189721] env[61906]: DEBUG nova.network.neutron [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.205225] env[61906]: DEBUG nova.network.neutron [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.442691] env[61906]: DEBUG nova.scheduler.client.report [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.510768] env[61906]: DEBUG nova.compute.manager [req-92150e8f-6552-4c88-8f2f-f3766dc36f90 req-315df43c-8443-4cd4-a9a4-84e3c1caaac1 service nova] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Received event network-changed-e107c8dc-cca7-47c3-b7ec-5d7c24a24f36 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 696.510948] env[61906]: DEBUG nova.compute.manager [req-92150e8f-6552-4c88-8f2f-f3766dc36f90 req-315df43c-8443-4cd4-a9a4-84e3c1caaac1 service nova] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Refreshing instance network info cache due to event network-changed-e107c8dc-cca7-47c3-b7ec-5d7c24a24f36. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 696.511182] env[61906]: DEBUG oslo_concurrency.lockutils [req-92150e8f-6552-4c88-8f2f-f3766dc36f90 req-315df43c-8443-4cd4-a9a4-84e3c1caaac1 service nova] Acquiring lock "refresh_cache-40dae609-e7a5-4af2-84d9-a53f93d95743" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.694024] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "refresh_cache-40dae609-e7a5-4af2-84d9-a53f93d95743" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.694024] env[61906]: DEBUG nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 696.694331] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 696.694547] env[61906]: DEBUG oslo_concurrency.lockutils [req-92150e8f-6552-4c88-8f2f-f3766dc36f90 req-315df43c-8443-4cd4-a9a4-84e3c1caaac1 service nova] Acquired lock "refresh_cache-40dae609-e7a5-4af2-84d9-a53f93d95743" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.694740] env[61906]: DEBUG nova.network.neutron [req-92150e8f-6552-4c88-8f2f-f3766dc36f90 req-315df43c-8443-4cd4-a9a4-84e3c1caaac1 service nova] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Refreshing network info cache for port e107c8dc-cca7-47c3-b7ec-5d7c24a24f36 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 696.695830] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cc7a7c5c-ab17-4be5-ae47-486e2a0bd2dd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.706930] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-182100fe-314c-4bbe-8394-dc1bc864df04 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.718738] env[61906]: DEBUG nova.network.neutron [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.731588] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 40dae609-e7a5-4af2-84d9-a53f93d95743 could not be found. [ 696.731892] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 696.732193] env[61906]: INFO nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Took 0.04 seconds to destroy the instance on the hypervisor. [ 696.732474] env[61906]: DEBUG oslo.service.loopingcall [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.733326] env[61906]: DEBUG nova.compute.manager [-] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.733415] env[61906]: DEBUG nova.network.neutron [-] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.750865] env[61906]: DEBUG nova.network.neutron [-] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.762421] env[61906]: DEBUG nova.network.neutron [-] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.948176] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.887s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.948807] env[61906]: ERROR nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 854dca8c-7653-47d3-a584-6fd29de10059, please check neutron logs for more information. [ 696.948807] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Traceback (most recent call last): [ 696.948807] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 696.948807] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] self.driver.spawn(context, instance, image_meta, [ 696.948807] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 696.948807] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.948807] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.948807] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] vm_ref = self.build_virtual_machine(instance, [ 696.948807] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.948807] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.948807] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] for vif in network_info: [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] return self._sync_wrapper(fn, *args, **kwargs) [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] self.wait() [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] self[:] = self._gt.wait() [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] return self._exit_event.wait() [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] current.throw(*self._exc) [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] result = function(*args, **kwargs) [ 696.949242] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 696.949738] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] return func(*args, **kwargs) [ 696.949738] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.949738] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] raise e [ 696.949738] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.949738] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] nwinfo = self.network_api.allocate_for_instance( [ 696.949738] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 696.949738] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] created_port_ids = self._update_ports_for_instance( [ 696.949738] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 696.949738] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] with excutils.save_and_reraise_exception(): [ 696.949738] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.949738] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] self.force_reraise() [ 696.949738] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.949738] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] raise self.value [ 696.950235] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 696.950235] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] updated_port = self._update_port( [ 696.950235] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.950235] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] _ensure_no_port_binding_failure(port) [ 696.950235] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.950235] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] raise exception.PortBindingFailed(port_id=port['id']) [ 696.950235] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] nova.exception.PortBindingFailed: Binding failed for port 854dca8c-7653-47d3-a584-6fd29de10059, please check neutron logs for more information. [ 696.950235] env[61906]: ERROR nova.compute.manager [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] [ 696.950235] env[61906]: DEBUG nova.compute.utils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Binding failed for port 854dca8c-7653-47d3-a584-6fd29de10059, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 696.951323] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.902s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.956522] env[61906]: DEBUG nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Build of instance 71fe23a8-1108-4781-9a35-0156cb9d584d was re-scheduled: Binding failed for port 854dca8c-7653-47d3-a584-6fd29de10059, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 696.956986] env[61906]: DEBUG nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 696.957674] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Acquiring lock "refresh_cache-71fe23a8-1108-4781-9a35-0156cb9d584d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.957674] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Acquired lock "refresh_cache-71fe23a8-1108-4781-9a35-0156cb9d584d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.957674] env[61906]: DEBUG nova.network.neutron [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.221442] env[61906]: INFO nova.compute.manager [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 42135fac-aef7-47d9-9afe-5ac2351bb07c] Took 1.04 seconds to deallocate network for instance. [ 697.224713] env[61906]: DEBUG nova.network.neutron [req-92150e8f-6552-4c88-8f2f-f3766dc36f90 req-315df43c-8443-4cd4-a9a4-84e3c1caaac1 service nova] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.265042] env[61906]: INFO nova.compute.manager [-] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Took 0.53 seconds to deallocate network for instance. [ 697.267954] env[61906]: DEBUG nova.compute.claims [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 697.268204] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.351615] env[61906]: DEBUG nova.network.neutron [req-92150e8f-6552-4c88-8f2f-f3766dc36f90 req-315df43c-8443-4cd4-a9a4-84e3c1caaac1 service nova] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.481752] env[61906]: DEBUG nova.network.neutron [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.562913] env[61906]: DEBUG nova.network.neutron [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.815610] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83b4446-785d-4a70-92df-2aed3af0aab3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.824802] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27369c67-797f-41c2-9305-5a55d821d338 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.855234] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8eaebb7-414c-4797-b354-131fea227e14 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.858045] env[61906]: DEBUG oslo_concurrency.lockutils [req-92150e8f-6552-4c88-8f2f-f3766dc36f90 req-315df43c-8443-4cd4-a9a4-84e3c1caaac1 service nova] Releasing lock "refresh_cache-40dae609-e7a5-4af2-84d9-a53f93d95743" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.858298] env[61906]: DEBUG nova.compute.manager [req-92150e8f-6552-4c88-8f2f-f3766dc36f90 req-315df43c-8443-4cd4-a9a4-84e3c1caaac1 service nova] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Received event network-vif-deleted-e107c8dc-cca7-47c3-b7ec-5d7c24a24f36 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 697.863596] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30c625c-1725-4471-b051-6f67d3ee7dcd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.876548] env[61906]: DEBUG nova.compute.provider_tree [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.065735] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Releasing lock "refresh_cache-71fe23a8-1108-4781-9a35-0156cb9d584d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.066113] env[61906]: DEBUG nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 698.066360] env[61906]: DEBUG nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.066982] env[61906]: DEBUG nova.network.neutron [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 698.095981] env[61906]: DEBUG nova.network.neutron [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.251111] env[61906]: INFO nova.scheduler.client.report [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Deleted allocations for instance 42135fac-aef7-47d9-9afe-5ac2351bb07c [ 698.380265] env[61906]: DEBUG nova.scheduler.client.report [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.598760] env[61906]: DEBUG nova.network.neutron [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.636092] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Acquiring lock "77791983-cef1-4f07-9b62-d52d335b08c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.636324] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Lock "77791983-cef1-4f07-9b62-d52d335b08c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.761674] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f527b7dd-359d-4df4-a452-8434e7134ed1 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "42135fac-aef7-47d9-9afe-5ac2351bb07c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.524s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.886264] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.935s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.886980] env[61906]: ERROR nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 96e6d7af-92f0-4732-be60-755f37c5d0f3, please check neutron logs for more information. [ 698.886980] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Traceback (most recent call last): [ 698.886980] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 698.886980] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] self.driver.spawn(context, instance, image_meta, [ 698.886980] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 698.886980] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] self._vmops.spawn(context, instance, image_meta, injected_files, [ 698.886980] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 698.886980] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] vm_ref = self.build_virtual_machine(instance, [ 698.886980] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 698.886980] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] vif_infos = vmwarevif.get_vif_info(self._session, [ 698.886980] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 698.887351] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] for vif in network_info: [ 698.887351] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 698.887351] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] return self._sync_wrapper(fn, *args, **kwargs) [ 698.887351] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 698.887351] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] self.wait() [ 698.887351] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 698.887351] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] self[:] = self._gt.wait() [ 698.887351] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 698.887351] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] return self._exit_event.wait() [ 698.887351] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 698.887351] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] result = hub.switch() [ 698.887351] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 698.887351] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] return self.greenlet.switch() [ 698.887699] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.887699] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] result = function(*args, **kwargs) [ 698.887699] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 698.887699] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] return func(*args, **kwargs) [ 698.887699] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 698.887699] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] raise e [ 698.887699] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.887699] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] nwinfo = self.network_api.allocate_for_instance( [ 698.887699] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.887699] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] created_port_ids = self._update_ports_for_instance( [ 698.887699] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.887699] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] with excutils.save_and_reraise_exception(): [ 698.887699] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.888063] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] self.force_reraise() [ 698.888063] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.888063] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] raise self.value [ 698.888063] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.888063] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] updated_port = self._update_port( [ 698.888063] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.888063] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] _ensure_no_port_binding_failure(port) [ 698.888063] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.888063] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] raise exception.PortBindingFailed(port_id=port['id']) [ 698.888063] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] nova.exception.PortBindingFailed: Binding failed for port 96e6d7af-92f0-4732-be60-755f37c5d0f3, please check neutron logs for more information. [ 698.888063] env[61906]: ERROR nova.compute.manager [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] [ 698.888403] env[61906]: DEBUG nova.compute.utils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Binding failed for port 96e6d7af-92f0-4732-be60-755f37c5d0f3, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 698.889206] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.474s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.890771] env[61906]: INFO nova.compute.claims [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 698.893611] env[61906]: DEBUG nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Build of instance 8d4c25cc-47d0-42e0-8fe9-49426175af68 was re-scheduled: Binding failed for port 96e6d7af-92f0-4732-be60-755f37c5d0f3, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 698.894138] env[61906]: DEBUG nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 698.894678] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquiring lock "refresh_cache-8d4c25cc-47d0-42e0-8fe9-49426175af68" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.894678] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquired lock "refresh_cache-8d4c25cc-47d0-42e0-8fe9-49426175af68" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.894678] env[61906]: DEBUG nova.network.neutron [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.101880] env[61906]: INFO nova.compute.manager [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] [instance: 71fe23a8-1108-4781-9a35-0156cb9d584d] Took 1.04 seconds to deallocate network for instance. [ 699.264743] env[61906]: DEBUG nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 699.418128] env[61906]: DEBUG nova.network.neutron [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.515139] env[61906]: DEBUG nova.network.neutron [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.798492] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.019034] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Releasing lock "refresh_cache-8d4c25cc-47d0-42e0-8fe9-49426175af68" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.019336] env[61906]: DEBUG nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 700.019478] env[61906]: DEBUG nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.019669] env[61906]: DEBUG nova.network.neutron [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.043632] env[61906]: DEBUG nova.network.neutron [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.134136] env[61906]: INFO nova.scheduler.client.report [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Deleted allocations for instance 71fe23a8-1108-4781-9a35-0156cb9d584d [ 700.216053] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9629d28a-2bb1-4565-95d6-98061e465e76 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.223747] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ee8fad-d471-4cb6-a821-7af38cbb6617 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.254367] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfd78b4-8fd1-4329-8574-191b11714919 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.261588] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baaef22a-3771-4612-b051-509463187002 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.274894] env[61906]: DEBUG nova.compute.provider_tree [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.546634] env[61906]: DEBUG nova.network.neutron [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.645476] env[61906]: DEBUG oslo_concurrency.lockutils [None req-31cc45e6-8a1f-41c3-8d45-8c89e82791c5 tempest-InstanceActionsV221TestJSON-2051165039 tempest-InstanceActionsV221TestJSON-2051165039-project-member] Lock "71fe23a8-1108-4781-9a35-0156cb9d584d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.341s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.779047] env[61906]: DEBUG nova.scheduler.client.report [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.048854] env[61906]: INFO nova.compute.manager [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 8d4c25cc-47d0-42e0-8fe9-49426175af68] Took 1.03 seconds to deallocate network for instance. [ 701.148970] env[61906]: DEBUG nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 701.283469] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.284052] env[61906]: DEBUG nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 701.286761] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.939s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.288246] env[61906]: INFO nova.compute.claims [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.676577] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.794011] env[61906]: DEBUG nova.compute.utils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 701.797236] env[61906]: DEBUG nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 701.797439] env[61906]: DEBUG nova.network.neutron [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 701.865013] env[61906]: DEBUG nova.policy [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9ca9e703d5d043cd82778754017b7201', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c54814414a124554b041331c6fb11a31', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 702.079079] env[61906]: INFO nova.scheduler.client.report [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Deleted allocations for instance 8d4c25cc-47d0-42e0-8fe9-49426175af68 [ 702.225294] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "7943ab1a-82df-48dd-874c-15dc015db51e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.225533] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "7943ab1a-82df-48dd-874c-15dc015db51e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.298448] env[61906]: DEBUG nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 702.393928] env[61906]: DEBUG nova.network.neutron [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Successfully created port: 83d41aa5-0562-48bd-b50b-eb5318c3fb13 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.589545] env[61906]: DEBUG oslo_concurrency.lockutils [None req-872230ac-82f7-4b85-9096-78b45054d094 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "8d4c25cc-47d0-42e0-8fe9-49426175af68" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.798s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.676609] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119c6e72-6aad-4eca-b992-8952248ddc6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.684638] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d902ce-82b9-4cfe-9a09-e3e54686463b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.718781] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cb7d7d4-73f5-4531-8836-3d4d15fa10bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.726184] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3870cb8e-ea96-46be-aa1e-02bedc4c917e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.741902] env[61906]: DEBUG nova.compute.provider_tree [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.099560] env[61906]: DEBUG nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 703.246197] env[61906]: DEBUG nova.scheduler.client.report [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.311178] env[61906]: DEBUG nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 703.338055] env[61906]: DEBUG nova.virt.hardware [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 703.338298] env[61906]: DEBUG nova.virt.hardware [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 703.339018] env[61906]: DEBUG nova.virt.hardware [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.339018] env[61906]: DEBUG nova.virt.hardware [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 703.339018] env[61906]: DEBUG nova.virt.hardware [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.339018] env[61906]: DEBUG nova.virt.hardware [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 703.339347] env[61906]: DEBUG nova.virt.hardware [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 703.339548] env[61906]: DEBUG nova.virt.hardware [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 703.339745] env[61906]: DEBUG nova.virt.hardware [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 703.339951] env[61906]: DEBUG nova.virt.hardware [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 703.340203] env[61906]: DEBUG nova.virt.hardware [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 703.342408] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-639b8ac5-653b-490b-8bde-b05651972dce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.349957] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e0f035-9888-4ff9-8067-61a6a90b7e30 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.626803] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.753565] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.754565] env[61906]: DEBUG nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 703.758155] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.752s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.861028] env[61906]: DEBUG nova.compute.manager [req-a5919b34-4069-4735-a51f-5ce0ec17e8ff req-deb240d8-6c81-4c00-a0ed-5ee535dde078 service nova] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Received event network-changed-83d41aa5-0562-48bd-b50b-eb5318c3fb13 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 703.861277] env[61906]: DEBUG nova.compute.manager [req-a5919b34-4069-4735-a51f-5ce0ec17e8ff req-deb240d8-6c81-4c00-a0ed-5ee535dde078 service nova] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Refreshing instance network info cache due to event network-changed-83d41aa5-0562-48bd-b50b-eb5318c3fb13. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 703.861498] env[61906]: DEBUG oslo_concurrency.lockutils [req-a5919b34-4069-4735-a51f-5ce0ec17e8ff req-deb240d8-6c81-4c00-a0ed-5ee535dde078 service nova] Acquiring lock "refresh_cache-59227a6d-e699-43e3-8f40-e8767ffdc938" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.861638] env[61906]: DEBUG oslo_concurrency.lockutils [req-a5919b34-4069-4735-a51f-5ce0ec17e8ff req-deb240d8-6c81-4c00-a0ed-5ee535dde078 service nova] Acquired lock "refresh_cache-59227a6d-e699-43e3-8f40-e8767ffdc938" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.861803] env[61906]: DEBUG nova.network.neutron [req-a5919b34-4069-4735-a51f-5ce0ec17e8ff req-deb240d8-6c81-4c00-a0ed-5ee535dde078 service nova] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Refreshing network info cache for port 83d41aa5-0562-48bd-b50b-eb5318c3fb13 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 704.169319] env[61906]: ERROR nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 83d41aa5-0562-48bd-b50b-eb5318c3fb13, please check neutron logs for more information. [ 704.169319] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 704.169319] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.169319] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 704.169319] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.169319] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 704.169319] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.169319] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 704.169319] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.169319] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 704.169319] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.169319] env[61906]: ERROR nova.compute.manager raise self.value [ 704.169319] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.169319] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 704.169319] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.169319] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 704.170128] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.170128] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 704.170128] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 83d41aa5-0562-48bd-b50b-eb5318c3fb13, please check neutron logs for more information. [ 704.170128] env[61906]: ERROR nova.compute.manager [ 704.170128] env[61906]: Traceback (most recent call last): [ 704.170128] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 704.170128] env[61906]: listener.cb(fileno) [ 704.170128] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.170128] env[61906]: result = function(*args, **kwargs) [ 704.170128] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 704.170128] env[61906]: return func(*args, **kwargs) [ 704.170128] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.170128] env[61906]: raise e [ 704.170128] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.170128] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 704.170128] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.170128] env[61906]: created_port_ids = self._update_ports_for_instance( [ 704.170128] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.170128] env[61906]: with excutils.save_and_reraise_exception(): [ 704.170128] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.170128] env[61906]: self.force_reraise() [ 704.170128] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.170128] env[61906]: raise self.value [ 704.170128] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.170128] env[61906]: updated_port = self._update_port( [ 704.170128] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.170128] env[61906]: _ensure_no_port_binding_failure(port) [ 704.170128] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.170128] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 704.171050] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 83d41aa5-0562-48bd-b50b-eb5318c3fb13, please check neutron logs for more information. [ 704.171050] env[61906]: Removing descriptor: 15 [ 704.171050] env[61906]: ERROR nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 83d41aa5-0562-48bd-b50b-eb5318c3fb13, please check neutron logs for more information. [ 704.171050] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Traceback (most recent call last): [ 704.171050] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 704.171050] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] yield resources [ 704.171050] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 704.171050] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] self.driver.spawn(context, instance, image_meta, [ 704.171050] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 704.171050] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.171050] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.171050] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] vm_ref = self.build_virtual_machine(instance, [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] for vif in network_info: [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] return self._sync_wrapper(fn, *args, **kwargs) [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] self.wait() [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] self[:] = self._gt.wait() [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] return self._exit_event.wait() [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 704.171545] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] result = hub.switch() [ 704.172436] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 704.172436] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] return self.greenlet.switch() [ 704.172436] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.172436] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] result = function(*args, **kwargs) [ 704.172436] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 704.172436] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] return func(*args, **kwargs) [ 704.172436] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.172436] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] raise e [ 704.172436] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.172436] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] nwinfo = self.network_api.allocate_for_instance( [ 704.172436] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.172436] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] created_port_ids = self._update_ports_for_instance( [ 704.172436] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.172881] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] with excutils.save_and_reraise_exception(): [ 704.172881] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.172881] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] self.force_reraise() [ 704.172881] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.172881] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] raise self.value [ 704.172881] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.172881] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] updated_port = self._update_port( [ 704.172881] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.172881] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] _ensure_no_port_binding_failure(port) [ 704.172881] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.172881] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] raise exception.PortBindingFailed(port_id=port['id']) [ 704.172881] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] nova.exception.PortBindingFailed: Binding failed for port 83d41aa5-0562-48bd-b50b-eb5318c3fb13, please check neutron logs for more information. [ 704.172881] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] [ 704.173302] env[61906]: INFO nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Terminating instance [ 704.175356] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Acquiring lock "refresh_cache-59227a6d-e699-43e3-8f40-e8767ffdc938" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.259503] env[61906]: DEBUG nova.compute.utils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.261075] env[61906]: DEBUG nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.261250] env[61906]: DEBUG nova.network.neutron [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 704.384348] env[61906]: DEBUG nova.network.neutron [req-a5919b34-4069-4735-a51f-5ce0ec17e8ff req-deb240d8-6c81-4c00-a0ed-5ee535dde078 service nova] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.453218] env[61906]: DEBUG nova.policy [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9dba4398586641c9adfe209ea853538f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17447d31611a46bbabae82a3ea2a2f22', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 704.513842] env[61906]: DEBUG nova.network.neutron [req-a5919b34-4069-4735-a51f-5ce0ec17e8ff req-deb240d8-6c81-4c00-a0ed-5ee535dde078 service nova] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.603854] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8791c8eb-0825-4111-9924-a132a8d349ca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.618322] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae85c1ce-b9e7-4e13-ad5c-703e44884a32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.648244] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730c8b8a-b357-4d6a-a569-bed0c986d5c7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.655438] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0fa2e6-2a4f-4af8-8461-09294007f858 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.668822] env[61906]: DEBUG nova.compute.provider_tree [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.766985] env[61906]: DEBUG nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 704.811168] env[61906]: DEBUG nova.network.neutron [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Successfully created port: 6452abb5-d435-41a5-a1eb-d41dcca5d2ab {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.017234] env[61906]: DEBUG oslo_concurrency.lockutils [req-a5919b34-4069-4735-a51f-5ce0ec17e8ff req-deb240d8-6c81-4c00-a0ed-5ee535dde078 service nova] Releasing lock "refresh_cache-59227a6d-e699-43e3-8f40-e8767ffdc938" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.017670] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Acquired lock "refresh_cache-59227a6d-e699-43e3-8f40-e8767ffdc938" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.017859] env[61906]: DEBUG nova.network.neutron [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.171940] env[61906]: DEBUG nova.scheduler.client.report [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.549477] env[61906]: DEBUG nova.network.neutron [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.681834] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.924s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.682455] env[61906]: ERROR nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3dae35ba-bed5-44cb-a046-47bcc1cdc551, please check neutron logs for more information. [ 705.682455] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Traceback (most recent call last): [ 705.682455] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.682455] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] self.driver.spawn(context, instance, image_meta, [ 705.682455] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 705.682455] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.682455] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.682455] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] vm_ref = self.build_virtual_machine(instance, [ 705.682455] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.682455] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.682455] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] for vif in network_info: [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] return self._sync_wrapper(fn, *args, **kwargs) [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] self.wait() [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] self[:] = self._gt.wait() [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] return self._exit_event.wait() [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] current.throw(*self._exc) [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] result = function(*args, **kwargs) [ 705.682860] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 705.683312] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] return func(*args, **kwargs) [ 705.683312] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.683312] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] raise e [ 705.683312] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.683312] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] nwinfo = self.network_api.allocate_for_instance( [ 705.683312] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.683312] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] created_port_ids = self._update_ports_for_instance( [ 705.683312] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.683312] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] with excutils.save_and_reraise_exception(): [ 705.683312] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.683312] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] self.force_reraise() [ 705.683312] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.683312] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] raise self.value [ 705.684062] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.684062] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] updated_port = self._update_port( [ 705.684062] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.684062] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] _ensure_no_port_binding_failure(port) [ 705.684062] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.684062] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] raise exception.PortBindingFailed(port_id=port['id']) [ 705.684062] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] nova.exception.PortBindingFailed: Binding failed for port 3dae35ba-bed5-44cb-a046-47bcc1cdc551, please check neutron logs for more information. [ 705.684062] env[61906]: ERROR nova.compute.manager [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] [ 705.684062] env[61906]: DEBUG nova.compute.utils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Binding failed for port 3dae35ba-bed5-44cb-a046-47bcc1cdc551, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 705.685561] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.045s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.690705] env[61906]: DEBUG nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Build of instance 2f5bd366-317e-4c18-956a-35d9cdbae9e2 was re-scheduled: Binding failed for port 3dae35ba-bed5-44cb-a046-47bcc1cdc551, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 705.690705] env[61906]: DEBUG nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 705.690705] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Acquiring lock "refresh_cache-2f5bd366-317e-4c18-956a-35d9cdbae9e2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.690705] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Acquired lock "refresh_cache-2f5bd366-317e-4c18-956a-35d9cdbae9e2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.690926] env[61906]: DEBUG nova.network.neutron [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.704190] env[61906]: DEBUG nova.network.neutron [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.777200] env[61906]: DEBUG nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 705.806346] env[61906]: DEBUG nova.virt.hardware [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 705.806927] env[61906]: DEBUG nova.virt.hardware [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 705.807214] env[61906]: DEBUG nova.virt.hardware [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 705.807534] env[61906]: DEBUG nova.virt.hardware [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 705.807790] env[61906]: DEBUG nova.virt.hardware [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 705.808050] env[61906]: DEBUG nova.virt.hardware [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 705.808359] env[61906]: DEBUG nova.virt.hardware [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 705.808625] env[61906]: DEBUG nova.virt.hardware [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 705.808927] env[61906]: DEBUG nova.virt.hardware [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 705.811847] env[61906]: DEBUG nova.virt.hardware [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 705.811847] env[61906]: DEBUG nova.virt.hardware [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 705.812552] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b68040b1-0001-432b-bbd2-de3a413c08a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.823711] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa5fe3c-f533-4716-bd0a-d6d336e52c5f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.898729] env[61906]: DEBUG nova.compute.manager [req-50f376f1-e148-4f74-b76c-455eed1f67cb req-1f975adf-5d17-4b90-b3ed-02e4577123f8 service nova] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Received event network-vif-deleted-83d41aa5-0562-48bd-b50b-eb5318c3fb13 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.898729] env[61906]: DEBUG nova.compute.manager [req-50f376f1-e148-4f74-b76c-455eed1f67cb req-1f975adf-5d17-4b90-b3ed-02e4577123f8 service nova] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Received event network-changed-6452abb5-d435-41a5-a1eb-d41dcca5d2ab {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.898729] env[61906]: DEBUG nova.compute.manager [req-50f376f1-e148-4f74-b76c-455eed1f67cb req-1f975adf-5d17-4b90-b3ed-02e4577123f8 service nova] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Refreshing instance network info cache due to event network-changed-6452abb5-d435-41a5-a1eb-d41dcca5d2ab. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 705.898893] env[61906]: DEBUG oslo_concurrency.lockutils [req-50f376f1-e148-4f74-b76c-455eed1f67cb req-1f975adf-5d17-4b90-b3ed-02e4577123f8 service nova] Acquiring lock "refresh_cache-2fdb9569-9e1a-44cd-9c2d-b493bc90ce81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.898944] env[61906]: DEBUG oslo_concurrency.lockutils [req-50f376f1-e148-4f74-b76c-455eed1f67cb req-1f975adf-5d17-4b90-b3ed-02e4577123f8 service nova] Acquired lock "refresh_cache-2fdb9569-9e1a-44cd-9c2d-b493bc90ce81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.899107] env[61906]: DEBUG nova.network.neutron [req-50f376f1-e148-4f74-b76c-455eed1f67cb req-1f975adf-5d17-4b90-b3ed-02e4577123f8 service nova] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Refreshing network info cache for port 6452abb5-d435-41a5-a1eb-d41dcca5d2ab {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 705.958617] env[61906]: ERROR nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6452abb5-d435-41a5-a1eb-d41dcca5d2ab, please check neutron logs for more information. [ 705.958617] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 705.958617] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.958617] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 705.958617] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.958617] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 705.958617] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.958617] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 705.958617] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.958617] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 705.958617] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.958617] env[61906]: ERROR nova.compute.manager raise self.value [ 705.958617] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.958617] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 705.958617] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.958617] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 705.959096] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.959096] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 705.959096] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6452abb5-d435-41a5-a1eb-d41dcca5d2ab, please check neutron logs for more information. [ 705.959096] env[61906]: ERROR nova.compute.manager [ 705.959096] env[61906]: Traceback (most recent call last): [ 705.959096] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 705.959096] env[61906]: listener.cb(fileno) [ 705.959096] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.959096] env[61906]: result = function(*args, **kwargs) [ 705.959096] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 705.959096] env[61906]: return func(*args, **kwargs) [ 705.959096] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.959096] env[61906]: raise e [ 705.959096] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.959096] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 705.959096] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.959096] env[61906]: created_port_ids = self._update_ports_for_instance( [ 705.959096] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.959096] env[61906]: with excutils.save_and_reraise_exception(): [ 705.959096] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.959096] env[61906]: self.force_reraise() [ 705.959096] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.959096] env[61906]: raise self.value [ 705.959096] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.959096] env[61906]: updated_port = self._update_port( [ 705.959096] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.959096] env[61906]: _ensure_no_port_binding_failure(port) [ 705.959096] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.959096] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 705.963372] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 6452abb5-d435-41a5-a1eb-d41dcca5d2ab, please check neutron logs for more information. [ 705.963372] env[61906]: Removing descriptor: 15 [ 705.963372] env[61906]: ERROR nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6452abb5-d435-41a5-a1eb-d41dcca5d2ab, please check neutron logs for more information. [ 705.963372] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Traceback (most recent call last): [ 705.963372] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 705.963372] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] yield resources [ 705.963372] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.963372] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] self.driver.spawn(context, instance, image_meta, [ 705.963372] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 705.963372] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.963372] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.963372] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] vm_ref = self.build_virtual_machine(instance, [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] for vif in network_info: [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] return self._sync_wrapper(fn, *args, **kwargs) [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] self.wait() [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] self[:] = self._gt.wait() [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] return self._exit_event.wait() [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 705.963870] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] result = hub.switch() [ 705.964310] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 705.964310] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] return self.greenlet.switch() [ 705.964310] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.964310] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] result = function(*args, **kwargs) [ 705.964310] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 705.964310] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] return func(*args, **kwargs) [ 705.964310] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.964310] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] raise e [ 705.964310] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.964310] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] nwinfo = self.network_api.allocate_for_instance( [ 705.964310] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.964310] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] created_port_ids = self._update_ports_for_instance( [ 705.964310] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.964704] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] with excutils.save_and_reraise_exception(): [ 705.964704] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.964704] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] self.force_reraise() [ 705.964704] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.964704] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] raise self.value [ 705.964704] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.964704] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] updated_port = self._update_port( [ 705.964704] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.964704] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] _ensure_no_port_binding_failure(port) [ 705.964704] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.964704] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] raise exception.PortBindingFailed(port_id=port['id']) [ 705.964704] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] nova.exception.PortBindingFailed: Binding failed for port 6452abb5-d435-41a5-a1eb-d41dcca5d2ab, please check neutron logs for more information. [ 705.964704] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] [ 705.965106] env[61906]: INFO nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Terminating instance [ 705.965106] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "refresh_cache-2fdb9569-9e1a-44cd-9c2d-b493bc90ce81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.205987] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Releasing lock "refresh_cache-59227a6d-e699-43e3-8f40-e8767ffdc938" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.206442] env[61906]: DEBUG nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 706.206636] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 706.206948] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f0f5f4d-7003-453c-9eaf-348f9fefd6ca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.218962] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c25dae7-05dc-4b19-9ad3-1f3e10ff02b7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.230917] env[61906]: DEBUG nova.network.neutron [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.247204] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 59227a6d-e699-43e3-8f40-e8767ffdc938 could not be found. [ 706.247204] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 706.247204] env[61906]: INFO nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Took 0.04 seconds to destroy the instance on the hypervisor. [ 706.247204] env[61906]: DEBUG oslo.service.loopingcall [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 706.247204] env[61906]: DEBUG nova.compute.manager [-] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.247397] env[61906]: DEBUG nova.network.neutron [-] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.276218] env[61906]: DEBUG nova.network.neutron [-] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.384425] env[61906]: DEBUG nova.network.neutron [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.432991] env[61906]: DEBUG nova.network.neutron [req-50f376f1-e148-4f74-b76c-455eed1f67cb req-1f975adf-5d17-4b90-b3ed-02e4577123f8 service nova] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.549602] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquiring lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.549602] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.557645] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c34117c-93a8-4486-aaa4-1c053da48d2a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.563753] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760fe982-ca49-4b8d-b703-1ff6171f1e8a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.594369] env[61906]: DEBUG nova.network.neutron [req-50f376f1-e148-4f74-b76c-455eed1f67cb req-1f975adf-5d17-4b90-b3ed-02e4577123f8 service nova] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.595951] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add4a3be-2eae-46cd-b91e-3bdacbb13a43 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.603715] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4309801f-a746-4c07-bf4a-fc03c4f230d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.619719] env[61906]: DEBUG nova.compute.provider_tree [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.780506] env[61906]: DEBUG nova.network.neutron [-] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.887564] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Releasing lock "refresh_cache-2f5bd366-317e-4c18-956a-35d9cdbae9e2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.887783] env[61906]: DEBUG nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 706.887977] env[61906]: DEBUG nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.888207] env[61906]: DEBUG nova.network.neutron [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.905186] env[61906]: DEBUG nova.network.neutron [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.100040] env[61906]: DEBUG oslo_concurrency.lockutils [req-50f376f1-e148-4f74-b76c-455eed1f67cb req-1f975adf-5d17-4b90-b3ed-02e4577123f8 service nova] Releasing lock "refresh_cache-2fdb9569-9e1a-44cd-9c2d-b493bc90ce81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.100354] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquired lock "refresh_cache-2fdb9569-9e1a-44cd-9c2d-b493bc90ce81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.100538] env[61906]: DEBUG nova.network.neutron [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.122299] env[61906]: DEBUG nova.scheduler.client.report [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.284086] env[61906]: INFO nova.compute.manager [-] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Took 1.04 seconds to deallocate network for instance. [ 707.286535] env[61906]: DEBUG nova.compute.claims [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 707.286623] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.409612] env[61906]: DEBUG nova.network.neutron [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.619849] env[61906]: DEBUG nova.network.neutron [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.628051] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.943s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.628723] env[61906]: ERROR nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e023640e-63ff-46e3-bc0c-7ccc3d59c920, please check neutron logs for more information. [ 707.628723] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Traceback (most recent call last): [ 707.628723] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.628723] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] self.driver.spawn(context, instance, image_meta, [ 707.628723] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 707.628723] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.628723] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.628723] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] vm_ref = self.build_virtual_machine(instance, [ 707.628723] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.628723] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.628723] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] for vif in network_info: [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] return self._sync_wrapper(fn, *args, **kwargs) [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] self.wait() [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] self[:] = self._gt.wait() [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] return self._exit_event.wait() [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] current.throw(*self._exc) [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] result = function(*args, **kwargs) [ 707.629093] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 707.629474] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] return func(*args, **kwargs) [ 707.629474] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.629474] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] raise e [ 707.629474] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.629474] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] nwinfo = self.network_api.allocate_for_instance( [ 707.629474] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.629474] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] created_port_ids = self._update_ports_for_instance( [ 707.629474] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.629474] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] with excutils.save_and_reraise_exception(): [ 707.629474] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.629474] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] self.force_reraise() [ 707.629474] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.629474] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] raise self.value [ 707.629832] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.629832] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] updated_port = self._update_port( [ 707.629832] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.629832] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] _ensure_no_port_binding_failure(port) [ 707.629832] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.629832] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] raise exception.PortBindingFailed(port_id=port['id']) [ 707.629832] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] nova.exception.PortBindingFailed: Binding failed for port e023640e-63ff-46e3-bc0c-7ccc3d59c920, please check neutron logs for more information. [ 707.629832] env[61906]: ERROR nova.compute.manager [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] [ 707.629832] env[61906]: DEBUG nova.compute.utils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Binding failed for port e023640e-63ff-46e3-bc0c-7ccc3d59c920, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 707.630580] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.566s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.632009] env[61906]: INFO nova.compute.claims [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.635423] env[61906]: DEBUG nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Build of instance cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b was re-scheduled: Binding failed for port e023640e-63ff-46e3-bc0c-7ccc3d59c920, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 707.636094] env[61906]: DEBUG nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 707.636394] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Acquiring lock "refresh_cache-cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.636568] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Acquired lock "refresh_cache-cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.636737] env[61906]: DEBUG nova.network.neutron [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.724210] env[61906]: DEBUG nova.network.neutron [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.912970] env[61906]: INFO nova.compute.manager [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] [instance: 2f5bd366-317e-4c18-956a-35d9cdbae9e2] Took 1.02 seconds to deallocate network for instance. [ 707.921497] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Acquiring lock "8ab6b473-d13c-4341-9789-992ac3aba6a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.921716] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Lock "8ab6b473-d13c-4341-9789-992ac3aba6a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.927968] env[61906]: DEBUG nova.compute.manager [req-833b0cf1-7c7b-4dac-9dbb-b41588df37c5 req-790fdf1a-e22e-4731-a94a-83b42a84312b service nova] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Received event network-vif-deleted-6452abb5-d435-41a5-a1eb-d41dcca5d2ab {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.164551] env[61906]: DEBUG nova.network.neutron [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.227756] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Releasing lock "refresh_cache-2fdb9569-9e1a-44cd-9c2d-b493bc90ce81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.228233] env[61906]: DEBUG nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 708.228458] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 708.228693] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de7be838-52ea-4f55-83e9-bcd3b0ba70ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.238057] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48dcb3bd-b7a8-4b1d-8ae8-5187568fa4fb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.249310] env[61906]: DEBUG nova.network.neutron [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.261459] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81 could not be found. [ 708.261608] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 708.261779] env[61906]: INFO nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Took 0.03 seconds to destroy the instance on the hypervisor. [ 708.262099] env[61906]: DEBUG oslo.service.loopingcall [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.262844] env[61906]: DEBUG nova.compute.manager [-] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.262944] env[61906]: DEBUG nova.network.neutron [-] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.278117] env[61906]: DEBUG nova.network.neutron [-] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.751888] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Releasing lock "refresh_cache-cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.752185] env[61906]: DEBUG nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 708.752399] env[61906]: DEBUG nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.752588] env[61906]: DEBUG nova.network.neutron [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.774505] env[61906]: DEBUG nova.network.neutron [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.782256] env[61906]: DEBUG nova.network.neutron [-] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.937452] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602d44ca-78cf-4a30-9239-db5e51f81587 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.941308] env[61906]: INFO nova.scheduler.client.report [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Deleted allocations for instance 2f5bd366-317e-4c18-956a-35d9cdbae9e2 [ 708.951797] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295c1282-f214-4eb9-934a-32f934854fb0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.986132] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a390a8f5-d2f2-4121-8fb9-2f109a3cab2f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.994281] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8d8bc3-9a4f-4c05-85fa-f19718017e46 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.009017] env[61906]: DEBUG nova.compute.provider_tree [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.278569] env[61906]: DEBUG nova.network.neutron [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.285500] env[61906]: INFO nova.compute.manager [-] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Took 1.02 seconds to deallocate network for instance. [ 709.288017] env[61906]: DEBUG nova.compute.claims [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 709.288125] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.450796] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cae228da-c8c6-4d6d-99e4-2d12b4271c0f tempest-ServerAddressesTestJSON-1725320268 tempest-ServerAddressesTestJSON-1725320268-project-member] Lock "2f5bd366-317e-4c18-956a-35d9cdbae9e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 143.112s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.511779] env[61906]: DEBUG nova.scheduler.client.report [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 709.782327] env[61906]: INFO nova.compute.manager [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] [instance: cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b] Took 1.03 seconds to deallocate network for instance. [ 709.953250] env[61906]: DEBUG nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 710.019022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.019022] env[61906]: DEBUG nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.020226] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.275s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.483364] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.525989] env[61906]: DEBUG nova.compute.utils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.531824] env[61906]: DEBUG nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 710.533250] env[61906]: DEBUG nova.network.neutron [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 710.599749] env[61906]: DEBUG nova.policy [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf278f5d85654c979056c6254ab65bac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '57747d3395a044ed8b09ce8216178698', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 710.813759] env[61906]: INFO nova.scheduler.client.report [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Deleted allocations for instance cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b [ 710.930161] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5044feb-a886-476a-b538-4652769aebb3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.938123] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce7152d-49c5-4784-98b3-5d7e6238a7d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.968066] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50cb3ae3-1706-49ce-9670-cf806abf26da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.975721] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c96d335-5caf-4535-91d4-d14e9ae47e3c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.991146] env[61906]: DEBUG nova.compute.provider_tree [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.038245] env[61906]: DEBUG nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 711.271691] env[61906]: DEBUG nova.network.neutron [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Successfully created port: 7b51d2d4-73a2-4d32-a5df-c0a27a33336c {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.333046] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cbbcbe13-18c5-42d9-8b85-aabe13fc6212 tempest-ServerDiagnosticsTest-409090786 tempest-ServerDiagnosticsTest-409090786-project-member] Lock "cb8d2bd7-ca6a-4465-8daa-9f15b13c9e2b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.342s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.497018] env[61906]: DEBUG nova.scheduler.client.report [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.837909] env[61906]: DEBUG nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 712.000229] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.980s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.000810] env[61906]: ERROR nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3a5c8162-640f-461e-9419-6593c9dad395, please check neutron logs for more information. [ 712.000810] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Traceback (most recent call last): [ 712.000810] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.000810] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] self.driver.spawn(context, instance, image_meta, [ 712.000810] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 712.000810] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.000810] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.000810] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] vm_ref = self.build_virtual_machine(instance, [ 712.000810] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.000810] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.000810] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] for vif in network_info: [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] return self._sync_wrapper(fn, *args, **kwargs) [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] self.wait() [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] self[:] = self._gt.wait() [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] return self._exit_event.wait() [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] current.throw(*self._exc) [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] result = function(*args, **kwargs) [ 712.001179] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.001586] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] return func(*args, **kwargs) [ 712.001586] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.001586] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] raise e [ 712.001586] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.001586] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] nwinfo = self.network_api.allocate_for_instance( [ 712.001586] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.001586] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] created_port_ids = self._update_ports_for_instance( [ 712.001586] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.001586] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] with excutils.save_and_reraise_exception(): [ 712.001586] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.001586] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] self.force_reraise() [ 712.001586] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.001586] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] raise self.value [ 712.001998] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.001998] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] updated_port = self._update_port( [ 712.001998] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.001998] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] _ensure_no_port_binding_failure(port) [ 712.001998] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.001998] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] raise exception.PortBindingFailed(port_id=port['id']) [ 712.001998] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] nova.exception.PortBindingFailed: Binding failed for port 3a5c8162-640f-461e-9419-6593c9dad395, please check neutron logs for more information. [ 712.001998] env[61906]: ERROR nova.compute.manager [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] [ 712.004254] env[61906]: DEBUG nova.compute.utils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Binding failed for port 3a5c8162-640f-461e-9419-6593c9dad395, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.005623] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.737s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.011314] env[61906]: DEBUG nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Build of instance 746ec539-5e13-4a9f-af73-26b175ad1e41 was re-scheduled: Binding failed for port 3a5c8162-640f-461e-9419-6593c9dad395, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 712.011314] env[61906]: DEBUG nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 712.012620] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquiring lock "refresh_cache-746ec539-5e13-4a9f-af73-26b175ad1e41" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.012859] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Acquired lock "refresh_cache-746ec539-5e13-4a9f-af73-26b175ad1e41" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.013030] env[61906]: DEBUG nova.network.neutron [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.044569] env[61906]: DEBUG nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.084740] env[61906]: DEBUG nova.virt.hardware [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.085008] env[61906]: DEBUG nova.virt.hardware [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.085174] env[61906]: DEBUG nova.virt.hardware [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.085353] env[61906]: DEBUG nova.virt.hardware [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.085498] env[61906]: DEBUG nova.virt.hardware [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.085643] env[61906]: DEBUG nova.virt.hardware [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.085849] env[61906]: DEBUG nova.virt.hardware [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.086017] env[61906]: DEBUG nova.virt.hardware [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.087047] env[61906]: DEBUG nova.virt.hardware [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.087236] env[61906]: DEBUG nova.virt.hardware [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.087671] env[61906]: DEBUG nova.virt.hardware [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.089144] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865662b3-269f-446e-8526-c5ff36ce6a14 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.098076] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad37c15b-e4bf-48ab-a5fb-a9c8c4be7acb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.369782] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.538163] env[61906]: DEBUG nova.network.neutron [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.624587] env[61906]: DEBUG nova.compute.manager [req-8f57f54b-89fb-45e1-8075-8ce886f284ce req-2f85d51f-bd94-40bc-9295-332dfadbbb27 service nova] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Received event network-changed-7b51d2d4-73a2-4d32-a5df-c0a27a33336c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 712.624861] env[61906]: DEBUG nova.compute.manager [req-8f57f54b-89fb-45e1-8075-8ce886f284ce req-2f85d51f-bd94-40bc-9295-332dfadbbb27 service nova] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Refreshing instance network info cache due to event network-changed-7b51d2d4-73a2-4d32-a5df-c0a27a33336c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 712.625155] env[61906]: DEBUG oslo_concurrency.lockutils [req-8f57f54b-89fb-45e1-8075-8ce886f284ce req-2f85d51f-bd94-40bc-9295-332dfadbbb27 service nova] Acquiring lock "refresh_cache-3c0602a1-ad7d-4f85-b4e0-d36565e28a38" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.625358] env[61906]: DEBUG oslo_concurrency.lockutils [req-8f57f54b-89fb-45e1-8075-8ce886f284ce req-2f85d51f-bd94-40bc-9295-332dfadbbb27 service nova] Acquired lock "refresh_cache-3c0602a1-ad7d-4f85-b4e0-d36565e28a38" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.625567] env[61906]: DEBUG nova.network.neutron [req-8f57f54b-89fb-45e1-8075-8ce886f284ce req-2f85d51f-bd94-40bc-9295-332dfadbbb27 service nova] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Refreshing network info cache for port 7b51d2d4-73a2-4d32-a5df-c0a27a33336c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 712.725890] env[61906]: DEBUG nova.network.neutron [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.826806] env[61906]: ERROR nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7b51d2d4-73a2-4d32-a5df-c0a27a33336c, please check neutron logs for more information. [ 712.826806] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 712.826806] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.826806] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 712.826806] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.826806] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 712.826806] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.826806] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 712.826806] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.826806] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 712.826806] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.826806] env[61906]: ERROR nova.compute.manager raise self.value [ 712.826806] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.826806] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 712.826806] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.826806] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 712.827315] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.827315] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 712.827315] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7b51d2d4-73a2-4d32-a5df-c0a27a33336c, please check neutron logs for more information. [ 712.827315] env[61906]: ERROR nova.compute.manager [ 712.827315] env[61906]: Traceback (most recent call last): [ 712.827315] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 712.827315] env[61906]: listener.cb(fileno) [ 712.827315] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.827315] env[61906]: result = function(*args, **kwargs) [ 712.827315] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.827315] env[61906]: return func(*args, **kwargs) [ 712.827315] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.827315] env[61906]: raise e [ 712.827315] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.827315] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 712.827315] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.827315] env[61906]: created_port_ids = self._update_ports_for_instance( [ 712.827315] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.827315] env[61906]: with excutils.save_and_reraise_exception(): [ 712.827315] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.827315] env[61906]: self.force_reraise() [ 712.827315] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.827315] env[61906]: raise self.value [ 712.827315] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.827315] env[61906]: updated_port = self._update_port( [ 712.827315] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.827315] env[61906]: _ensure_no_port_binding_failure(port) [ 712.827315] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.827315] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 712.829686] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 7b51d2d4-73a2-4d32-a5df-c0a27a33336c, please check neutron logs for more information. [ 712.829686] env[61906]: Removing descriptor: 15 [ 712.829686] env[61906]: ERROR nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7b51d2d4-73a2-4d32-a5df-c0a27a33336c, please check neutron logs for more information. [ 712.829686] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Traceback (most recent call last): [ 712.829686] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 712.829686] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] yield resources [ 712.829686] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.829686] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] self.driver.spawn(context, instance, image_meta, [ 712.829686] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 712.829686] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.829686] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.829686] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] vm_ref = self.build_virtual_machine(instance, [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] for vif in network_info: [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] return self._sync_wrapper(fn, *args, **kwargs) [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] self.wait() [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] self[:] = self._gt.wait() [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] return self._exit_event.wait() [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 712.830798] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] result = hub.switch() [ 712.831350] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 712.831350] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] return self.greenlet.switch() [ 712.831350] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.831350] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] result = function(*args, **kwargs) [ 712.831350] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.831350] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] return func(*args, **kwargs) [ 712.831350] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.831350] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] raise e [ 712.831350] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.831350] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] nwinfo = self.network_api.allocate_for_instance( [ 712.831350] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.831350] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] created_port_ids = self._update_ports_for_instance( [ 712.831350] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.832309] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] with excutils.save_and_reraise_exception(): [ 712.832309] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.832309] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] self.force_reraise() [ 712.832309] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.832309] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] raise self.value [ 712.832309] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.832309] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] updated_port = self._update_port( [ 712.832309] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.832309] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] _ensure_no_port_binding_failure(port) [ 712.832309] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.832309] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] raise exception.PortBindingFailed(port_id=port['id']) [ 712.832309] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] nova.exception.PortBindingFailed: Binding failed for port 7b51d2d4-73a2-4d32-a5df-c0a27a33336c, please check neutron logs for more information. [ 712.832309] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] [ 712.833022] env[61906]: INFO nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Terminating instance [ 712.834581] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Acquiring lock "refresh_cache-3c0602a1-ad7d-4f85-b4e0-d36565e28a38" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.857670] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e5afeb-4f01-4190-93c9-6b8498141720 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.865517] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-368db0f1-dff9-4f93-ad5c-1f2189cf8624 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.898962] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e82ecd3-63ec-44cf-b266-54c97a73ad58 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.906808] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a582b3e0-d6ed-40ec-a8d4-c398943d93ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.920395] env[61906]: DEBUG nova.compute.provider_tree [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.156075] env[61906]: DEBUG nova.network.neutron [req-8f57f54b-89fb-45e1-8075-8ce886f284ce req-2f85d51f-bd94-40bc-9295-332dfadbbb27 service nova] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.229414] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Releasing lock "refresh_cache-746ec539-5e13-4a9f-af73-26b175ad1e41" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.229572] env[61906]: DEBUG nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 713.230150] env[61906]: DEBUG nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.230150] env[61906]: DEBUG nova.network.neutron [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.249974] env[61906]: DEBUG nova.network.neutron [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.271859] env[61906]: DEBUG nova.network.neutron [req-8f57f54b-89fb-45e1-8075-8ce886f284ce req-2f85d51f-bd94-40bc-9295-332dfadbbb27 service nova] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.426189] env[61906]: DEBUG nova.scheduler.client.report [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 713.754044] env[61906]: DEBUG nova.network.neutron [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.774410] env[61906]: DEBUG oslo_concurrency.lockutils [req-8f57f54b-89fb-45e1-8075-8ce886f284ce req-2f85d51f-bd94-40bc-9295-332dfadbbb27 service nova] Releasing lock "refresh_cache-3c0602a1-ad7d-4f85-b4e0-d36565e28a38" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.774829] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Acquired lock "refresh_cache-3c0602a1-ad7d-4f85-b4e0-d36565e28a38" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.775028] env[61906]: DEBUG nova.network.neutron [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.931814] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.926s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.932565] env[61906]: ERROR nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e107c8dc-cca7-47c3-b7ec-5d7c24a24f36, please check neutron logs for more information. [ 713.932565] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Traceback (most recent call last): [ 713.932565] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 713.932565] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] self.driver.spawn(context, instance, image_meta, [ 713.932565] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 713.932565] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] self._vmops.spawn(context, instance, image_meta, injected_files, [ 713.932565] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 713.932565] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] vm_ref = self.build_virtual_machine(instance, [ 713.932565] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 713.932565] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] vif_infos = vmwarevif.get_vif_info(self._session, [ 713.932565] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 713.932898] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] for vif in network_info: [ 713.932898] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 713.932898] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] return self._sync_wrapper(fn, *args, **kwargs) [ 713.932898] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 713.932898] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] self.wait() [ 713.932898] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 713.932898] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] self[:] = self._gt.wait() [ 713.932898] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 713.932898] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] return self._exit_event.wait() [ 713.932898] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 713.932898] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] result = hub.switch() [ 713.932898] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 713.932898] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] return self.greenlet.switch() [ 713.933274] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.933274] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] result = function(*args, **kwargs) [ 713.933274] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 713.933274] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] return func(*args, **kwargs) [ 713.933274] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.933274] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] raise e [ 713.933274] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.933274] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] nwinfo = self.network_api.allocate_for_instance( [ 713.933274] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.933274] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] created_port_ids = self._update_ports_for_instance( [ 713.933274] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.933274] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] with excutils.save_and_reraise_exception(): [ 713.933274] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.933655] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] self.force_reraise() [ 713.933655] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.933655] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] raise self.value [ 713.933655] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.933655] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] updated_port = self._update_port( [ 713.933655] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.933655] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] _ensure_no_port_binding_failure(port) [ 713.933655] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.933655] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] raise exception.PortBindingFailed(port_id=port['id']) [ 713.933655] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] nova.exception.PortBindingFailed: Binding failed for port e107c8dc-cca7-47c3-b7ec-5d7c24a24f36, please check neutron logs for more information. [ 713.933655] env[61906]: ERROR nova.compute.manager [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] [ 713.934026] env[61906]: DEBUG nova.compute.utils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Binding failed for port e107c8dc-cca7-47c3-b7ec-5d7c24a24f36, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 713.934671] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.136s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.936143] env[61906]: INFO nova.compute.claims [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.938975] env[61906]: DEBUG nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Build of instance 40dae609-e7a5-4af2-84d9-a53f93d95743 was re-scheduled: Binding failed for port e107c8dc-cca7-47c3-b7ec-5d7c24a24f36, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 713.939383] env[61906]: DEBUG nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 713.939609] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "refresh_cache-40dae609-e7a5-4af2-84d9-a53f93d95743" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.939757] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "refresh_cache-40dae609-e7a5-4af2-84d9-a53f93d95743" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.939915] env[61906]: DEBUG nova.network.neutron [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 714.256684] env[61906]: INFO nova.compute.manager [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] [instance: 746ec539-5e13-4a9f-af73-26b175ad1e41] Took 1.03 seconds to deallocate network for instance. [ 714.295704] env[61906]: DEBUG nova.network.neutron [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.375783] env[61906]: DEBUG nova.network.neutron [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.458357] env[61906]: DEBUG nova.network.neutron [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.538700] env[61906]: DEBUG nova.network.neutron [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.646474] env[61906]: DEBUG nova.compute.manager [req-d9bdebeb-825c-4f10-9663-5476b4631291 req-e13179d5-f0b1-4359-95de-f9629a147fc6 service nova] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Received event network-vif-deleted-7b51d2d4-73a2-4d32-a5df-c0a27a33336c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.877789] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Releasing lock "refresh_cache-3c0602a1-ad7d-4f85-b4e0-d36565e28a38" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.878287] env[61906]: DEBUG nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 714.878478] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 714.880055] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-35022162-5a61-4e57-9028-ccfbf5d0de67 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.888850] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9a6987-a0ea-4fc1-b593-8634c03124d9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.909886] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3c0602a1-ad7d-4f85-b4e0-d36565e28a38 could not be found. [ 714.910110] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 714.910295] env[61906]: INFO nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Took 0.03 seconds to destroy the instance on the hypervisor. [ 714.910526] env[61906]: DEBUG oslo.service.loopingcall [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 714.910737] env[61906]: DEBUG nova.compute.manager [-] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 714.910829] env[61906]: DEBUG nova.network.neutron [-] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.924982] env[61906]: DEBUG nova.network.neutron [-] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.041217] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "refresh_cache-40dae609-e7a5-4af2-84d9-a53f93d95743" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.041470] env[61906]: DEBUG nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 715.041641] env[61906]: DEBUG nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.041813] env[61906]: DEBUG nova.network.neutron [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 715.056852] env[61906]: DEBUG nova.network.neutron [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.214321] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-526b62b9-4d09-4eb8-b7e0-484d0751705b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.221952] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9292d71d-66e4-415f-bf78-37acd0ccac71 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.252034] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92481c9b-8b4d-4618-9088-5ba7dc359388 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.258217] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984b8d1d-64ff-4dbc-aa50-6c522b90c69d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.273433] env[61906]: DEBUG nova.compute.provider_tree [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.293580] env[61906]: INFO nova.scheduler.client.report [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Deleted allocations for instance 746ec539-5e13-4a9f-af73-26b175ad1e41 [ 715.427073] env[61906]: DEBUG nova.network.neutron [-] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.553988] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 715.554279] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 715.560533] env[61906]: DEBUG nova.network.neutron [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.777630] env[61906]: DEBUG nova.scheduler.client.report [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.801101] env[61906]: DEBUG oslo_concurrency.lockutils [None req-eb634769-4cfc-4753-bb87-8d825d062307 tempest-VolumesAdminNegativeTest-159563549 tempest-VolumesAdminNegativeTest-159563549-project-member] Lock "746ec539-5e13-4a9f-af73-26b175ad1e41" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.774s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.929925] env[61906]: INFO nova.compute.manager [-] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Took 1.02 seconds to deallocate network for instance. [ 715.932914] env[61906]: DEBUG nova.compute.claims [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 715.933197] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.061140] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.061140] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 716.061140] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 716.065829] env[61906]: INFO nova.compute.manager [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 40dae609-e7a5-4af2-84d9-a53f93d95743] Took 1.02 seconds to deallocate network for instance. [ 716.282633] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.348s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.283088] env[61906]: DEBUG nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 716.285694] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.609s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.287113] env[61906]: INFO nova.compute.claims [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 716.303901] env[61906]: DEBUG nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 716.564480] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 716.564480] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 716.564480] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 716.564480] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 716.564693] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Didn't find any instances for network info cache update. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 716.564780] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.564966] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.565089] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.565267] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.565413] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.565556] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.565695] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 716.565869] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.791643] env[61906]: DEBUG nova.compute.utils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 716.794999] env[61906]: DEBUG nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 716.795185] env[61906]: DEBUG nova.network.neutron [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 716.824481] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.880145] env[61906]: DEBUG nova.policy [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '882091491d004ebfa1017749b405dfeb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2594b33f677a47e4ad5db08f6469d744', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.069347] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.107894] env[61906]: INFO nova.scheduler.client.report [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleted allocations for instance 40dae609-e7a5-4af2-84d9-a53f93d95743 [ 717.296127] env[61906]: DEBUG nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 717.331291] env[61906]: DEBUG nova.network.neutron [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Successfully created port: 909abf08-0182-4d09-a329-3d5a76441acb {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.589787] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7fb2d8f-fc25-431b-a657-b0e9c7e21d3a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.599037] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748eb65d-c83b-4256-adb7-7153b88de52f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.633726] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f48885f-5fb5-46bc-aac6-7a77fefcaabf tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "40dae609-e7a5-4af2-84d9-a53f93d95743" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.758s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.635729] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de16bba8-97d3-4143-b219-2db98de199b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.648241] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381bba44-b436-41e8-ad42-87c1909d9bf9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.665672] env[61906]: DEBUG nova.compute.provider_tree [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.135203] env[61906]: DEBUG nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 718.163841] env[61906]: DEBUG nova.scheduler.client.report [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.309193] env[61906]: DEBUG nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 718.367023] env[61906]: DEBUG nova.virt.hardware [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 718.367023] env[61906]: DEBUG nova.virt.hardware [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 718.367023] env[61906]: DEBUG nova.virt.hardware [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.367241] env[61906]: DEBUG nova.virt.hardware [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 718.367499] env[61906]: DEBUG nova.virt.hardware [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.367600] env[61906]: DEBUG nova.virt.hardware [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 718.368457] env[61906]: DEBUG nova.virt.hardware [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 718.368457] env[61906]: DEBUG nova.virt.hardware [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 718.368633] env[61906]: DEBUG nova.virt.hardware [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 718.368863] env[61906]: DEBUG nova.virt.hardware [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 718.369121] env[61906]: DEBUG nova.virt.hardware [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.370340] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034b1213-157f-455c-b078-ca5a0c932ebc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.381181] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c31ce143-dbca-4d69-9457-b74d9a6da094 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.521322] env[61906]: DEBUG nova.compute.manager [req-6920eb54-6691-4476-90ec-96d8aa68a6fa req-a224be3b-3c20-4dd5-84b1-fb57b2d0670d service nova] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Received event network-changed-909abf08-0182-4d09-a329-3d5a76441acb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.522949] env[61906]: DEBUG nova.compute.manager [req-6920eb54-6691-4476-90ec-96d8aa68a6fa req-a224be3b-3c20-4dd5-84b1-fb57b2d0670d service nova] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Refreshing instance network info cache due to event network-changed-909abf08-0182-4d09-a329-3d5a76441acb. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 718.522949] env[61906]: DEBUG oslo_concurrency.lockutils [req-6920eb54-6691-4476-90ec-96d8aa68a6fa req-a224be3b-3c20-4dd5-84b1-fb57b2d0670d service nova] Acquiring lock "refresh_cache-4a2ed7bd-eb49-49a0-ba81-3c29baea6533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.522949] env[61906]: DEBUG oslo_concurrency.lockutils [req-6920eb54-6691-4476-90ec-96d8aa68a6fa req-a224be3b-3c20-4dd5-84b1-fb57b2d0670d service nova] Acquired lock "refresh_cache-4a2ed7bd-eb49-49a0-ba81-3c29baea6533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.522949] env[61906]: DEBUG nova.network.neutron [req-6920eb54-6691-4476-90ec-96d8aa68a6fa req-a224be3b-3c20-4dd5-84b1-fb57b2d0670d service nova] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Refreshing network info cache for port 909abf08-0182-4d09-a329-3d5a76441acb {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 718.657891] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.670117] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.670741] env[61906]: DEBUG nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 718.673630] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.047s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.675971] env[61906]: INFO nova.compute.claims [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.872822] env[61906]: ERROR nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 909abf08-0182-4d09-a329-3d5a76441acb, please check neutron logs for more information. [ 718.872822] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 718.872822] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.872822] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 718.872822] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.872822] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 718.872822] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.872822] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 718.872822] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.872822] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 718.872822] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.872822] env[61906]: ERROR nova.compute.manager raise self.value [ 718.872822] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.872822] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 718.872822] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.872822] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 718.873377] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.873377] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 718.873377] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 909abf08-0182-4d09-a329-3d5a76441acb, please check neutron logs for more information. [ 718.873377] env[61906]: ERROR nova.compute.manager [ 718.873377] env[61906]: Traceback (most recent call last): [ 718.873377] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 718.873377] env[61906]: listener.cb(fileno) [ 718.873377] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.873377] env[61906]: result = function(*args, **kwargs) [ 718.873377] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.873377] env[61906]: return func(*args, **kwargs) [ 718.873377] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.873377] env[61906]: raise e [ 718.873377] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.873377] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 718.873377] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.873377] env[61906]: created_port_ids = self._update_ports_for_instance( [ 718.873377] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.873377] env[61906]: with excutils.save_and_reraise_exception(): [ 718.873377] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.873377] env[61906]: self.force_reraise() [ 718.873377] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.873377] env[61906]: raise self.value [ 718.873377] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.873377] env[61906]: updated_port = self._update_port( [ 718.873377] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.873377] env[61906]: _ensure_no_port_binding_failure(port) [ 718.873377] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.873377] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 718.874323] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 909abf08-0182-4d09-a329-3d5a76441acb, please check neutron logs for more information. [ 718.874323] env[61906]: Removing descriptor: 15 [ 718.874323] env[61906]: ERROR nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 909abf08-0182-4d09-a329-3d5a76441acb, please check neutron logs for more information. [ 718.874323] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Traceback (most recent call last): [ 718.874323] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 718.874323] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] yield resources [ 718.874323] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.874323] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] self.driver.spawn(context, instance, image_meta, [ 718.874323] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 718.874323] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.874323] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.874323] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] vm_ref = self.build_virtual_machine(instance, [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] for vif in network_info: [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] return self._sync_wrapper(fn, *args, **kwargs) [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] self.wait() [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] self[:] = self._gt.wait() [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] return self._exit_event.wait() [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 718.874743] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] result = hub.switch() [ 718.875211] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 718.875211] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] return self.greenlet.switch() [ 718.875211] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.875211] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] result = function(*args, **kwargs) [ 718.875211] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.875211] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] return func(*args, **kwargs) [ 718.875211] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.875211] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] raise e [ 718.875211] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.875211] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] nwinfo = self.network_api.allocate_for_instance( [ 718.875211] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.875211] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] created_port_ids = self._update_ports_for_instance( [ 718.875211] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.875637] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] with excutils.save_and_reraise_exception(): [ 718.875637] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.875637] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] self.force_reraise() [ 718.875637] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.875637] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] raise self.value [ 718.875637] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.875637] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] updated_port = self._update_port( [ 718.875637] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.875637] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] _ensure_no_port_binding_failure(port) [ 718.875637] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.875637] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] raise exception.PortBindingFailed(port_id=port['id']) [ 718.875637] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] nova.exception.PortBindingFailed: Binding failed for port 909abf08-0182-4d09-a329-3d5a76441acb, please check neutron logs for more information. [ 718.875637] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] [ 718.876129] env[61906]: INFO nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Terminating instance [ 718.876426] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquiring lock "refresh_cache-4a2ed7bd-eb49-49a0-ba81-3c29baea6533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.991900] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.991900] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.044455] env[61906]: DEBUG nova.network.neutron [req-6920eb54-6691-4476-90ec-96d8aa68a6fa req-a224be3b-3c20-4dd5-84b1-fb57b2d0670d service nova] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.128067] env[61906]: DEBUG nova.network.neutron [req-6920eb54-6691-4476-90ec-96d8aa68a6fa req-a224be3b-3c20-4dd5-84b1-fb57b2d0670d service nova] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.181188] env[61906]: DEBUG nova.compute.utils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 719.185025] env[61906]: DEBUG nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 719.185240] env[61906]: DEBUG nova.network.neutron [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 719.285053] env[61906]: DEBUG nova.policy [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '882091491d004ebfa1017749b405dfeb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2594b33f677a47e4ad5db08f6469d744', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 719.631720] env[61906]: DEBUG oslo_concurrency.lockutils [req-6920eb54-6691-4476-90ec-96d8aa68a6fa req-a224be3b-3c20-4dd5-84b1-fb57b2d0670d service nova] Releasing lock "refresh_cache-4a2ed7bd-eb49-49a0-ba81-3c29baea6533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.631720] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquired lock "refresh_cache-4a2ed7bd-eb49-49a0-ba81-3c29baea6533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.632132] env[61906]: DEBUG nova.network.neutron [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.685574] env[61906]: DEBUG nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 719.787545] env[61906]: DEBUG nova.network.neutron [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Successfully created port: 3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.984204] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1780d5-ded2-4c7c-90b0-6c17375f1d72 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.990761] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8233a2b9-d9ed-439b-a4bc-947962128729 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.025285] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab3e252e-a065-4c8d-8565-59e9d9ab49ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.036031] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65234e76-4a26-4a8b-b47b-9e9eb70ad6a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.049274] env[61906]: DEBUG nova.compute.provider_tree [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.151157] env[61906]: DEBUG nova.network.neutron [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.210485] env[61906]: DEBUG nova.network.neutron [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.551872] env[61906]: DEBUG nova.scheduler.client.report [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.558411] env[61906]: DEBUG nova.compute.manager [req-836cdbb5-d5a7-4582-a5f1-49c53811cfa5 req-43defb7d-2ee4-4d3e-b8cb-e3fc49ff604b service nova] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Received event network-vif-deleted-909abf08-0182-4d09-a329-3d5a76441acb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.698429] env[61906]: DEBUG nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 720.714439] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Releasing lock "refresh_cache-4a2ed7bd-eb49-49a0-ba81-3c29baea6533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.714439] env[61906]: DEBUG nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 720.714439] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.714715] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e292f27f-7a1a-49ae-b8aa-ef4dc71f214d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.727673] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00cc261-b272-4ef7-a4db-d22864142e3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.739939] env[61906]: DEBUG nova.virt.hardware [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.739939] env[61906]: DEBUG nova.virt.hardware [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.740134] env[61906]: DEBUG nova.virt.hardware [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.740822] env[61906]: DEBUG nova.virt.hardware [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.740822] env[61906]: DEBUG nova.virt.hardware [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.740822] env[61906]: DEBUG nova.virt.hardware [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.740822] env[61906]: DEBUG nova.virt.hardware [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.740822] env[61906]: DEBUG nova.virt.hardware [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.741068] env[61906]: DEBUG nova.virt.hardware [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.741292] env[61906]: DEBUG nova.virt.hardware [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.741345] env[61906]: DEBUG nova.virt.hardware [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.742122] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12514e1-223e-4b1c-91a1-241e44065746 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.750768] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb4f5359-7b3c-40c3-997f-e9476b83ba7f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.757443] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4a2ed7bd-eb49-49a0-ba81-3c29baea6533 could not be found. [ 720.757618] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 720.757795] env[61906]: INFO nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Took 0.04 seconds to destroy the instance on the hypervisor. [ 720.758051] env[61906]: DEBUG oslo.service.loopingcall [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.758611] env[61906]: DEBUG nova.compute.manager [-] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.758700] env[61906]: DEBUG nova.network.neutron [-] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.790430] env[61906]: DEBUG nova.network.neutron [-] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.057404] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.058105] env[61906]: DEBUG nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 721.062938] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.776s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.193521] env[61906]: ERROR nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8, please check neutron logs for more information. [ 721.193521] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 721.193521] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.193521] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 721.193521] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.193521] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 721.193521] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.193521] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 721.193521] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.193521] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 721.193521] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.193521] env[61906]: ERROR nova.compute.manager raise self.value [ 721.193521] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.193521] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 721.193521] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.193521] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 721.194107] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.194107] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 721.194107] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8, please check neutron logs for more information. [ 721.194107] env[61906]: ERROR nova.compute.manager [ 721.194107] env[61906]: Traceback (most recent call last): [ 721.194107] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 721.194107] env[61906]: listener.cb(fileno) [ 721.194107] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.194107] env[61906]: result = function(*args, **kwargs) [ 721.194107] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 721.194107] env[61906]: return func(*args, **kwargs) [ 721.194107] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.194107] env[61906]: raise e [ 721.194107] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.194107] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 721.194107] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.194107] env[61906]: created_port_ids = self._update_ports_for_instance( [ 721.194107] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.194107] env[61906]: with excutils.save_and_reraise_exception(): [ 721.194107] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.194107] env[61906]: self.force_reraise() [ 721.194107] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.194107] env[61906]: raise self.value [ 721.194107] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.194107] env[61906]: updated_port = self._update_port( [ 721.194107] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.194107] env[61906]: _ensure_no_port_binding_failure(port) [ 721.194107] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.194107] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 721.195124] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8, please check neutron logs for more information. [ 721.195124] env[61906]: Removing descriptor: 15 [ 721.195124] env[61906]: ERROR nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8, please check neutron logs for more information. [ 721.195124] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Traceback (most recent call last): [ 721.195124] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 721.195124] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] yield resources [ 721.195124] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.195124] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] self.driver.spawn(context, instance, image_meta, [ 721.195124] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 721.195124] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.195124] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.195124] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] vm_ref = self.build_virtual_machine(instance, [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] for vif in network_info: [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] return self._sync_wrapper(fn, *args, **kwargs) [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] self.wait() [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] self[:] = self._gt.wait() [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] return self._exit_event.wait() [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 721.195535] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] result = hub.switch() [ 721.196029] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 721.196029] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] return self.greenlet.switch() [ 721.196029] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.196029] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] result = function(*args, **kwargs) [ 721.196029] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 721.196029] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] return func(*args, **kwargs) [ 721.196029] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.196029] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] raise e [ 721.196029] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.196029] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] nwinfo = self.network_api.allocate_for_instance( [ 721.196029] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.196029] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] created_port_ids = self._update_ports_for_instance( [ 721.196029] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.196459] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] with excutils.save_and_reraise_exception(): [ 721.196459] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.196459] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] self.force_reraise() [ 721.196459] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.196459] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] raise self.value [ 721.196459] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.196459] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] updated_port = self._update_port( [ 721.196459] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.196459] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] _ensure_no_port_binding_failure(port) [ 721.196459] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.196459] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] raise exception.PortBindingFailed(port_id=port['id']) [ 721.196459] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] nova.exception.PortBindingFailed: Binding failed for port 3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8, please check neutron logs for more information. [ 721.196459] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] [ 721.197411] env[61906]: INFO nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Terminating instance [ 721.197411] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquiring lock "refresh_cache-0dca3209-d37b-4536-bbf9-72f0418cef8e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.197411] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquired lock "refresh_cache-0dca3209-d37b-4536-bbf9-72f0418cef8e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.197411] env[61906]: DEBUG nova.network.neutron [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.293511] env[61906]: DEBUG nova.network.neutron [-] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.568652] env[61906]: DEBUG nova.compute.utils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.572693] env[61906]: DEBUG nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 721.572693] env[61906]: DEBUG nova.network.neutron [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 721.645077] env[61906]: DEBUG nova.policy [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d5405d35a72490dbe5ce244c88346b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08b1224bc5494ab6a02e84eed71005c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 721.715075] env[61906]: DEBUG nova.network.neutron [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.775394] env[61906]: DEBUG nova.network.neutron [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.796615] env[61906]: INFO nova.compute.manager [-] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Took 1.04 seconds to deallocate network for instance. [ 721.801478] env[61906]: DEBUG nova.compute.claims [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.801670] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.864501] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08d18ae-091b-4b34-9d0e-78ec646761c9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.872930] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584f33ce-ec5c-4e3e-bd9d-7d6cd2d2ee7f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.904051] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa33e58-fef7-4499-8bb0-cd31e119d155 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.912147] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c83247c7-f17c-4824-9e70-bb98c4252a1e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.929019] env[61906]: DEBUG nova.compute.provider_tree [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.073898] env[61906]: DEBUG nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 722.266903] env[61906]: DEBUG nova.network.neutron [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Successfully created port: 95c18951-8f02-411b-9709-3412b832e328 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.278610] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Releasing lock "refresh_cache-0dca3209-d37b-4536-bbf9-72f0418cef8e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.279050] env[61906]: DEBUG nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 722.279256] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 722.279548] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4c12d4b4-7108-404d-be51-1a99e51c9798 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.289791] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c507a2f-b264-476e-b711-f3716b2bbc9b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.309565] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0dca3209-d37b-4536-bbf9-72f0418cef8e could not be found. [ 722.309884] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 722.309961] env[61906]: INFO nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 722.310203] env[61906]: DEBUG oslo.service.loopingcall [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.310414] env[61906]: DEBUG nova.compute.manager [-] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.310507] env[61906]: DEBUG nova.network.neutron [-] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.324921] env[61906]: DEBUG nova.network.neutron [-] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.433657] env[61906]: DEBUG nova.scheduler.client.report [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.590414] env[61906]: DEBUG nova.compute.manager [req-92546e66-41e2-423c-ae17-6d9cc351b434 req-4d32313f-d3c1-4ba6-a3da-988556bfb40c service nova] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Received event network-changed-3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.590621] env[61906]: DEBUG nova.compute.manager [req-92546e66-41e2-423c-ae17-6d9cc351b434 req-4d32313f-d3c1-4ba6-a3da-988556bfb40c service nova] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Refreshing instance network info cache due to event network-changed-3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 722.590821] env[61906]: DEBUG oslo_concurrency.lockutils [req-92546e66-41e2-423c-ae17-6d9cc351b434 req-4d32313f-d3c1-4ba6-a3da-988556bfb40c service nova] Acquiring lock "refresh_cache-0dca3209-d37b-4536-bbf9-72f0418cef8e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.590996] env[61906]: DEBUG oslo_concurrency.lockutils [req-92546e66-41e2-423c-ae17-6d9cc351b434 req-4d32313f-d3c1-4ba6-a3da-988556bfb40c service nova] Acquired lock "refresh_cache-0dca3209-d37b-4536-bbf9-72f0418cef8e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.591127] env[61906]: DEBUG nova.network.neutron [req-92546e66-41e2-423c-ae17-6d9cc351b434 req-4d32313f-d3c1-4ba6-a3da-988556bfb40c service nova] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Refreshing network info cache for port 3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 722.827918] env[61906]: DEBUG nova.network.neutron [-] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.944020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.880s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.944020] env[61906]: ERROR nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 83d41aa5-0562-48bd-b50b-eb5318c3fb13, please check neutron logs for more information. [ 722.944020] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Traceback (most recent call last): [ 722.944020] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 722.944020] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] self.driver.spawn(context, instance, image_meta, [ 722.944020] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 722.944020] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.944020] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.944020] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] vm_ref = self.build_virtual_machine(instance, [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] for vif in network_info: [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] return self._sync_wrapper(fn, *args, **kwargs) [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] self.wait() [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] self[:] = self._gt.wait() [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] return self._exit_event.wait() [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 722.944438] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] result = hub.switch() [ 722.944872] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 722.944872] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] return self.greenlet.switch() [ 722.944872] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.944872] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] result = function(*args, **kwargs) [ 722.944872] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 722.944872] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] return func(*args, **kwargs) [ 722.944872] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.944872] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] raise e [ 722.944872] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.944872] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] nwinfo = self.network_api.allocate_for_instance( [ 722.944872] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.944872] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] created_port_ids = self._update_ports_for_instance( [ 722.944872] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.945301] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] with excutils.save_and_reraise_exception(): [ 722.945301] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.945301] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] self.force_reraise() [ 722.945301] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.945301] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] raise self.value [ 722.945301] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.945301] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] updated_port = self._update_port( [ 722.945301] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.945301] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] _ensure_no_port_binding_failure(port) [ 722.945301] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.945301] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] raise exception.PortBindingFailed(port_id=port['id']) [ 722.945301] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] nova.exception.PortBindingFailed: Binding failed for port 83d41aa5-0562-48bd-b50b-eb5318c3fb13, please check neutron logs for more information. [ 722.945301] env[61906]: ERROR nova.compute.manager [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] [ 722.945658] env[61906]: DEBUG nova.compute.utils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Binding failed for port 83d41aa5-0562-48bd-b50b-eb5318c3fb13, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 722.945658] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.657s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.948496] env[61906]: DEBUG nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Build of instance 59227a6d-e699-43e3-8f40-e8767ffdc938 was re-scheduled: Binding failed for port 83d41aa5-0562-48bd-b50b-eb5318c3fb13, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 722.950012] env[61906]: DEBUG nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 722.950012] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Acquiring lock "refresh_cache-59227a6d-e699-43e3-8f40-e8767ffdc938" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.950012] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Acquired lock "refresh_cache-59227a6d-e699-43e3-8f40-e8767ffdc938" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.950012] env[61906]: DEBUG nova.network.neutron [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 723.088522] env[61906]: DEBUG nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.117477] env[61906]: DEBUG nova.virt.hardware [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.117738] env[61906]: DEBUG nova.virt.hardware [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.117897] env[61906]: DEBUG nova.virt.hardware [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.118180] env[61906]: DEBUG nova.virt.hardware [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.118339] env[61906]: DEBUG nova.virt.hardware [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.118486] env[61906]: DEBUG nova.virt.hardware [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.118703] env[61906]: DEBUG nova.virt.hardware [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.118868] env[61906]: DEBUG nova.virt.hardware [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.119245] env[61906]: DEBUG nova.virt.hardware [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.119479] env[61906]: DEBUG nova.virt.hardware [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.119661] env[61906]: DEBUG nova.virt.hardware [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.121241] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deeaa2a7-eeaf-4c43-9fef-da60801c13b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.124860] env[61906]: DEBUG nova.network.neutron [req-92546e66-41e2-423c-ae17-6d9cc351b434 req-4d32313f-d3c1-4ba6-a3da-988556bfb40c service nova] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.134428] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3930c6e9-173e-4cac-8739-c8cfaf04bf32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.206516] env[61906]: DEBUG nova.network.neutron [req-92546e66-41e2-423c-ae17-6d9cc351b434 req-4d32313f-d3c1-4ba6-a3da-988556bfb40c service nova] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.308185] env[61906]: ERROR nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 95c18951-8f02-411b-9709-3412b832e328, please check neutron logs for more information. [ 723.308185] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 723.308185] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.308185] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 723.308185] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.308185] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 723.308185] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.308185] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 723.308185] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.308185] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 723.308185] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.308185] env[61906]: ERROR nova.compute.manager raise self.value [ 723.308185] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.308185] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 723.308185] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.308185] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 723.308765] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.308765] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 723.308765] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 95c18951-8f02-411b-9709-3412b832e328, please check neutron logs for more information. [ 723.308765] env[61906]: ERROR nova.compute.manager [ 723.308765] env[61906]: Traceback (most recent call last): [ 723.308765] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 723.308765] env[61906]: listener.cb(fileno) [ 723.308765] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.308765] env[61906]: result = function(*args, **kwargs) [ 723.308765] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 723.308765] env[61906]: return func(*args, **kwargs) [ 723.308765] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.308765] env[61906]: raise e [ 723.308765] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.308765] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 723.308765] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.308765] env[61906]: created_port_ids = self._update_ports_for_instance( [ 723.308765] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.308765] env[61906]: with excutils.save_and_reraise_exception(): [ 723.308765] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.308765] env[61906]: self.force_reraise() [ 723.308765] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.308765] env[61906]: raise self.value [ 723.308765] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.308765] env[61906]: updated_port = self._update_port( [ 723.308765] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.308765] env[61906]: _ensure_no_port_binding_failure(port) [ 723.308765] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.308765] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 723.309560] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 95c18951-8f02-411b-9709-3412b832e328, please check neutron logs for more information. [ 723.309560] env[61906]: Removing descriptor: 15 [ 723.309560] env[61906]: ERROR nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 95c18951-8f02-411b-9709-3412b832e328, please check neutron logs for more information. [ 723.309560] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Traceback (most recent call last): [ 723.309560] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 723.309560] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] yield resources [ 723.309560] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 723.309560] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] self.driver.spawn(context, instance, image_meta, [ 723.309560] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 723.309560] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.309560] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.309560] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] vm_ref = self.build_virtual_machine(instance, [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] for vif in network_info: [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] return self._sync_wrapper(fn, *args, **kwargs) [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] self.wait() [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] self[:] = self._gt.wait() [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] return self._exit_event.wait() [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 723.310173] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] result = hub.switch() [ 723.310585] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 723.310585] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] return self.greenlet.switch() [ 723.310585] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.310585] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] result = function(*args, **kwargs) [ 723.310585] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 723.310585] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] return func(*args, **kwargs) [ 723.310585] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.310585] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] raise e [ 723.310585] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.310585] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] nwinfo = self.network_api.allocate_for_instance( [ 723.310585] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.310585] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] created_port_ids = self._update_ports_for_instance( [ 723.310585] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.310968] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] with excutils.save_and_reraise_exception(): [ 723.310968] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.310968] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] self.force_reraise() [ 723.310968] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.310968] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] raise self.value [ 723.310968] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.310968] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] updated_port = self._update_port( [ 723.310968] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.310968] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] _ensure_no_port_binding_failure(port) [ 723.310968] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.310968] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] raise exception.PortBindingFailed(port_id=port['id']) [ 723.310968] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] nova.exception.PortBindingFailed: Binding failed for port 95c18951-8f02-411b-9709-3412b832e328, please check neutron logs for more information. [ 723.310968] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] [ 723.312045] env[61906]: INFO nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Terminating instance [ 723.312045] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Acquiring lock "refresh_cache-b1ec5d55-96a6-42eb-ba54-c4a3904e22cd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.312045] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Acquired lock "refresh_cache-b1ec5d55-96a6-42eb-ba54-c4a3904e22cd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.312359] env[61906]: DEBUG nova.network.neutron [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 723.330159] env[61906]: INFO nova.compute.manager [-] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Took 1.02 seconds to deallocate network for instance. [ 723.332337] env[61906]: DEBUG nova.compute.claims [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 723.332512] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.482707] env[61906]: DEBUG nova.network.neutron [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.559079] env[61906]: DEBUG nova.network.neutron [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.709309] env[61906]: DEBUG oslo_concurrency.lockutils [req-92546e66-41e2-423c-ae17-6d9cc351b434 req-4d32313f-d3c1-4ba6-a3da-988556bfb40c service nova] Releasing lock "refresh_cache-0dca3209-d37b-4536-bbf9-72f0418cef8e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.709527] env[61906]: DEBUG nova.compute.manager [req-92546e66-41e2-423c-ae17-6d9cc351b434 req-4d32313f-d3c1-4ba6-a3da-988556bfb40c service nova] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Received event network-vif-deleted-3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 723.758209] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830bf056-8bbd-4930-9c71-2f0fa431c8eb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.765521] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3808e32-3a67-4879-838e-eafaf9b4d3f1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.795227] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdee0615-f44c-42b3-912a-ca944cec75e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.802433] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5bfda5-5a51-4d9c-8043-9441e47beeef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.817715] env[61906]: DEBUG nova.compute.provider_tree [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.840752] env[61906]: DEBUG nova.network.neutron [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.892969] env[61906]: DEBUG nova.network.neutron [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.064055] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Releasing lock "refresh_cache-59227a6d-e699-43e3-8f40-e8767ffdc938" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.064055] env[61906]: DEBUG nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 724.064055] env[61906]: DEBUG nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.064055] env[61906]: DEBUG nova.network.neutron [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 724.078286] env[61906]: DEBUG nova.network.neutron [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.321080] env[61906]: DEBUG nova.scheduler.client.report [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.398601] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Releasing lock "refresh_cache-b1ec5d55-96a6-42eb-ba54-c4a3904e22cd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.398601] env[61906]: DEBUG nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 724.398601] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 724.398601] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41a87a64-e392-45ca-9a29-38dbbda171c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.408098] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e85f8d-549c-4ffb-8642-7a3e3936c7a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.430608] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b1ec5d55-96a6-42eb-ba54-c4a3904e22cd could not be found. [ 724.430846] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 724.431077] env[61906]: INFO nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Took 0.03 seconds to destroy the instance on the hypervisor. [ 724.431291] env[61906]: DEBUG oslo.service.loopingcall [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.431496] env[61906]: DEBUG nova.compute.manager [-] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.431589] env[61906]: DEBUG nova.network.neutron [-] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 724.450232] env[61906]: DEBUG nova.network.neutron [-] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.584019] env[61906]: DEBUG nova.network.neutron [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.618578] env[61906]: DEBUG nova.compute.manager [req-58797cfe-13e5-4069-beab-5a62c7a2babc req-292171cb-36a6-419f-99c3-1079b6229916 service nova] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Received event network-changed-95c18951-8f02-411b-9709-3412b832e328 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 724.618780] env[61906]: DEBUG nova.compute.manager [req-58797cfe-13e5-4069-beab-5a62c7a2babc req-292171cb-36a6-419f-99c3-1079b6229916 service nova] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Refreshing instance network info cache due to event network-changed-95c18951-8f02-411b-9709-3412b832e328. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 724.619016] env[61906]: DEBUG oslo_concurrency.lockutils [req-58797cfe-13e5-4069-beab-5a62c7a2babc req-292171cb-36a6-419f-99c3-1079b6229916 service nova] Acquiring lock "refresh_cache-b1ec5d55-96a6-42eb-ba54-c4a3904e22cd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.619155] env[61906]: DEBUG oslo_concurrency.lockutils [req-58797cfe-13e5-4069-beab-5a62c7a2babc req-292171cb-36a6-419f-99c3-1079b6229916 service nova] Acquired lock "refresh_cache-b1ec5d55-96a6-42eb-ba54-c4a3904e22cd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.619315] env[61906]: DEBUG nova.network.neutron [req-58797cfe-13e5-4069-beab-5a62c7a2babc req-292171cb-36a6-419f-99c3-1079b6229916 service nova] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Refreshing network info cache for port 95c18951-8f02-411b-9709-3412b832e328 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 724.830021] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.881s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.830021] env[61906]: ERROR nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6452abb5-d435-41a5-a1eb-d41dcca5d2ab, please check neutron logs for more information. [ 724.830021] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Traceback (most recent call last): [ 724.830021] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.830021] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] self.driver.spawn(context, instance, image_meta, [ 724.830021] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 724.830021] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.830021] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.830021] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] vm_ref = self.build_virtual_machine(instance, [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] for vif in network_info: [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] return self._sync_wrapper(fn, *args, **kwargs) [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] self.wait() [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] self[:] = self._gt.wait() [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] return self._exit_event.wait() [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 724.830544] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] result = hub.switch() [ 724.830948] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 724.830948] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] return self.greenlet.switch() [ 724.830948] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.830948] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] result = function(*args, **kwargs) [ 724.830948] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.830948] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] return func(*args, **kwargs) [ 724.830948] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.830948] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] raise e [ 724.830948] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.830948] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] nwinfo = self.network_api.allocate_for_instance( [ 724.830948] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.830948] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] created_port_ids = self._update_ports_for_instance( [ 724.830948] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.831320] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] with excutils.save_and_reraise_exception(): [ 724.831320] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.831320] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] self.force_reraise() [ 724.831320] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.831320] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] raise self.value [ 724.831320] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.831320] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] updated_port = self._update_port( [ 724.831320] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.831320] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] _ensure_no_port_binding_failure(port) [ 724.831320] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.831320] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] raise exception.PortBindingFailed(port_id=port['id']) [ 724.831320] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] nova.exception.PortBindingFailed: Binding failed for port 6452abb5-d435-41a5-a1eb-d41dcca5d2ab, please check neutron logs for more information. [ 724.831320] env[61906]: ERROR nova.compute.manager [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] [ 724.831792] env[61906]: DEBUG nova.compute.utils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Binding failed for port 6452abb5-d435-41a5-a1eb-d41dcca5d2ab, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 724.831792] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.346s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.831792] env[61906]: INFO nova.compute.claims [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.833551] env[61906]: DEBUG nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Build of instance 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81 was re-scheduled: Binding failed for port 6452abb5-d435-41a5-a1eb-d41dcca5d2ab, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 724.834096] env[61906]: DEBUG nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 724.834312] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "refresh_cache-2fdb9569-9e1a-44cd-9c2d-b493bc90ce81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.834459] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquired lock "refresh_cache-2fdb9569-9e1a-44cd-9c2d-b493bc90ce81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.834662] env[61906]: DEBUG nova.network.neutron [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.953293] env[61906]: DEBUG nova.network.neutron [-] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.086151] env[61906]: INFO nova.compute.manager [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] [instance: 59227a6d-e699-43e3-8f40-e8767ffdc938] Took 1.02 seconds to deallocate network for instance. [ 725.136755] env[61906]: DEBUG nova.network.neutron [req-58797cfe-13e5-4069-beab-5a62c7a2babc req-292171cb-36a6-419f-99c3-1079b6229916 service nova] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.202092] env[61906]: DEBUG nova.network.neutron [req-58797cfe-13e5-4069-beab-5a62c7a2babc req-292171cb-36a6-419f-99c3-1079b6229916 service nova] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.354023] env[61906]: DEBUG nova.network.neutron [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.456025] env[61906]: DEBUG nova.network.neutron [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.457480] env[61906]: INFO nova.compute.manager [-] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Took 1.03 seconds to deallocate network for instance. [ 725.459911] env[61906]: DEBUG nova.compute.claims [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 725.460680] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.705314] env[61906]: DEBUG oslo_concurrency.lockutils [req-58797cfe-13e5-4069-beab-5a62c7a2babc req-292171cb-36a6-419f-99c3-1079b6229916 service nova] Releasing lock "refresh_cache-b1ec5d55-96a6-42eb-ba54-c4a3904e22cd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.705574] env[61906]: DEBUG nova.compute.manager [req-58797cfe-13e5-4069-beab-5a62c7a2babc req-292171cb-36a6-419f-99c3-1079b6229916 service nova] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Received event network-vif-deleted-95c18951-8f02-411b-9709-3412b832e328 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.959603] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Releasing lock "refresh_cache-2fdb9569-9e1a-44cd-9c2d-b493bc90ce81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.959603] env[61906]: DEBUG nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 725.959763] env[61906]: DEBUG nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.959924] env[61906]: DEBUG nova.network.neutron [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.982339] env[61906]: DEBUG nova.network.neutron [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.085995] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "5adaa660-b736-4c11-9141-846cf475ccd5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.086267] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "5adaa660-b736-4c11-9141-846cf475ccd5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.120552] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46977e8c-59a5-4419-a405-1c258055422d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.124073] env[61906]: INFO nova.scheduler.client.report [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Deleted allocations for instance 59227a6d-e699-43e3-8f40-e8767ffdc938 [ 726.138352] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa32344-3af2-4ecf-8bd6-97119c25123f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.169825] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f659b02c-811d-4181-9c55-a457dd243d32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.177799] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca71040-392f-47fb-97ce-33ff140ac106 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.191746] env[61906]: DEBUG nova.compute.provider_tree [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.486486] env[61906]: DEBUG nova.network.neutron [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.632809] env[61906]: DEBUG oslo_concurrency.lockutils [None req-11555628-2f90-42ec-863e-3a9ee7b15c58 tempest-ServerActionsTestOtherA-1135015098 tempest-ServerActionsTestOtherA-1135015098-project-member] Lock "59227a6d-e699-43e3-8f40-e8767ffdc938" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.402s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.695086] env[61906]: DEBUG nova.scheduler.client.report [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.990729] env[61906]: INFO nova.compute.manager [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81] Took 1.03 seconds to deallocate network for instance. [ 727.135300] env[61906]: DEBUG nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 727.200372] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.201791] env[61906]: DEBUG nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 727.203643] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.834s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.205421] env[61906]: INFO nova.compute.claims [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.660182] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.712219] env[61906]: DEBUG nova.compute.utils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 727.712219] env[61906]: DEBUG nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 727.712219] env[61906]: DEBUG nova.network.neutron [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 727.759665] env[61906]: DEBUG nova.policy [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd4614eac1b543bbae9e6cbdfc737ece', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1bb8b80700b2403f91072183df36a35b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 728.030197] env[61906]: INFO nova.scheduler.client.report [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Deleted allocations for instance 2fdb9569-9e1a-44cd-9c2d-b493bc90ce81 [ 728.037207] env[61906]: DEBUG nova.network.neutron [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Successfully created port: d4ce567a-fd81-426f-af76-0a6fe4e164c1 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 728.215063] env[61906]: DEBUG nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 728.538116] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2074bb31-2313-4707-be37-11809b9c49f1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.545579] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-444be35c-44c3-4953-b4b6-b3ddcc311b4e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.549027] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1edcb7c5-c585-4c56-b338-2a3f5168207c tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "2fdb9569-9e1a-44cd-9c2d-b493bc90ce81" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.611s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.582582] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900f3893-5909-40d9-b128-3ebe940a0e31 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.590983] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8a5fdc-d180-4d03-a911-ff2443cfdd48 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.610056] env[61906]: DEBUG nova.compute.provider_tree [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.008903] env[61906]: DEBUG nova.compute.manager [req-d6a28d91-93d7-4aa7-91b8-9dcb3745bc42 req-dbd9e399-3a3e-4a44-976c-15da767d95e1 service nova] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Received event network-changed-d4ce567a-fd81-426f-af76-0a6fe4e164c1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.009194] env[61906]: DEBUG nova.compute.manager [req-d6a28d91-93d7-4aa7-91b8-9dcb3745bc42 req-dbd9e399-3a3e-4a44-976c-15da767d95e1 service nova] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Refreshing instance network info cache due to event network-changed-d4ce567a-fd81-426f-af76-0a6fe4e164c1. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 729.009378] env[61906]: DEBUG oslo_concurrency.lockutils [req-d6a28d91-93d7-4aa7-91b8-9dcb3745bc42 req-dbd9e399-3a3e-4a44-976c-15da767d95e1 service nova] Acquiring lock "refresh_cache-83c86407-16b6-4025-9568-0c9c73d20b10" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.009602] env[61906]: DEBUG oslo_concurrency.lockutils [req-d6a28d91-93d7-4aa7-91b8-9dcb3745bc42 req-dbd9e399-3a3e-4a44-976c-15da767d95e1 service nova] Acquired lock "refresh_cache-83c86407-16b6-4025-9568-0c9c73d20b10" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.009602] env[61906]: DEBUG nova.network.neutron [req-d6a28d91-93d7-4aa7-91b8-9dcb3745bc42 req-dbd9e399-3a3e-4a44-976c-15da767d95e1 service nova] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Refreshing network info cache for port d4ce567a-fd81-426f-af76-0a6fe4e164c1 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 729.051618] env[61906]: DEBUG nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 729.114298] env[61906]: DEBUG nova.scheduler.client.report [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.122184] env[61906]: ERROR nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d4ce567a-fd81-426f-af76-0a6fe4e164c1, please check neutron logs for more information. [ 729.122184] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 729.122184] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.122184] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 729.122184] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.122184] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 729.122184] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.122184] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 729.122184] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.122184] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 729.122184] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.122184] env[61906]: ERROR nova.compute.manager raise self.value [ 729.122184] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.122184] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 729.122184] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.122184] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 729.122729] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.122729] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 729.122729] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d4ce567a-fd81-426f-af76-0a6fe4e164c1, please check neutron logs for more information. [ 729.122729] env[61906]: ERROR nova.compute.manager [ 729.122729] env[61906]: Traceback (most recent call last): [ 729.122729] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 729.122729] env[61906]: listener.cb(fileno) [ 729.122729] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.122729] env[61906]: result = function(*args, **kwargs) [ 729.122729] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.122729] env[61906]: return func(*args, **kwargs) [ 729.122729] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.122729] env[61906]: raise e [ 729.122729] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.122729] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 729.122729] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.122729] env[61906]: created_port_ids = self._update_ports_for_instance( [ 729.122729] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.122729] env[61906]: with excutils.save_and_reraise_exception(): [ 729.122729] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.122729] env[61906]: self.force_reraise() [ 729.122729] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.122729] env[61906]: raise self.value [ 729.122729] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.122729] env[61906]: updated_port = self._update_port( [ 729.122729] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.122729] env[61906]: _ensure_no_port_binding_failure(port) [ 729.122729] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.122729] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 729.123622] env[61906]: nova.exception.PortBindingFailed: Binding failed for port d4ce567a-fd81-426f-af76-0a6fe4e164c1, please check neutron logs for more information. [ 729.123622] env[61906]: Removing descriptor: 15 [ 729.226744] env[61906]: DEBUG nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 729.260859] env[61906]: DEBUG nova.virt.hardware [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 729.260859] env[61906]: DEBUG nova.virt.hardware [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 729.260859] env[61906]: DEBUG nova.virt.hardware [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 729.261095] env[61906]: DEBUG nova.virt.hardware [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 729.261249] env[61906]: DEBUG nova.virt.hardware [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 729.261400] env[61906]: DEBUG nova.virt.hardware [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 729.261609] env[61906]: DEBUG nova.virt.hardware [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 729.261765] env[61906]: DEBUG nova.virt.hardware [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 729.261931] env[61906]: DEBUG nova.virt.hardware [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 729.262229] env[61906]: DEBUG nova.virt.hardware [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 729.262296] env[61906]: DEBUG nova.virt.hardware [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.263222] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c0f31e-5183-452b-9333-4549a2a17e6e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.271728] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715cc8cb-1b2a-41e9-8272-70c82531f74f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.285990] env[61906]: ERROR nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d4ce567a-fd81-426f-af76-0a6fe4e164c1, please check neutron logs for more information. [ 729.285990] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Traceback (most recent call last): [ 729.285990] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 729.285990] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] yield resources [ 729.285990] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 729.285990] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] self.driver.spawn(context, instance, image_meta, [ 729.285990] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 729.285990] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] self._vmops.spawn(context, instance, image_meta, injected_files, [ 729.285990] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 729.285990] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] vm_ref = self.build_virtual_machine(instance, [ 729.285990] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 729.286408] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] vif_infos = vmwarevif.get_vif_info(self._session, [ 729.286408] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 729.286408] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] for vif in network_info: [ 729.286408] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 729.286408] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] return self._sync_wrapper(fn, *args, **kwargs) [ 729.286408] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 729.286408] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] self.wait() [ 729.286408] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 729.286408] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] self[:] = self._gt.wait() [ 729.286408] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 729.286408] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] return self._exit_event.wait() [ 729.286408] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 729.286408] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] current.throw(*self._exc) [ 729.286780] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.286780] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] result = function(*args, **kwargs) [ 729.286780] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.286780] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] return func(*args, **kwargs) [ 729.286780] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.286780] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] raise e [ 729.286780] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.286780] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] nwinfo = self.network_api.allocate_for_instance( [ 729.286780] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.286780] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] created_port_ids = self._update_ports_for_instance( [ 729.286780] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.286780] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] with excutils.save_and_reraise_exception(): [ 729.286780] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.287257] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] self.force_reraise() [ 729.287257] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.287257] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] raise self.value [ 729.287257] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.287257] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] updated_port = self._update_port( [ 729.287257] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.287257] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] _ensure_no_port_binding_failure(port) [ 729.287257] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.287257] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] raise exception.PortBindingFailed(port_id=port['id']) [ 729.287257] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] nova.exception.PortBindingFailed: Binding failed for port d4ce567a-fd81-426f-af76-0a6fe4e164c1, please check neutron logs for more information. [ 729.287257] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] [ 729.287257] env[61906]: INFO nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Terminating instance [ 729.288546] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Acquiring lock "refresh_cache-83c86407-16b6-4025-9568-0c9c73d20b10" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.528726] env[61906]: DEBUG nova.network.neutron [req-d6a28d91-93d7-4aa7-91b8-9dcb3745bc42 req-dbd9e399-3a3e-4a44-976c-15da767d95e1 service nova] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.577454] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.614594] env[61906]: DEBUG nova.network.neutron [req-d6a28d91-93d7-4aa7-91b8-9dcb3745bc42 req-dbd9e399-3a3e-4a44-976c-15da767d95e1 service nova] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.620300] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.620801] env[61906]: DEBUG nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 729.623501] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.690s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.116571] env[61906]: DEBUG oslo_concurrency.lockutils [req-d6a28d91-93d7-4aa7-91b8-9dcb3745bc42 req-dbd9e399-3a3e-4a44-976c-15da767d95e1 service nova] Releasing lock "refresh_cache-83c86407-16b6-4025-9568-0c9c73d20b10" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.118021] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Acquired lock "refresh_cache-83c86407-16b6-4025-9568-0c9c73d20b10" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.118147] env[61906]: DEBUG nova.network.neutron [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.127836] env[61906]: DEBUG nova.compute.utils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 730.134476] env[61906]: DEBUG nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 730.134996] env[61906]: DEBUG nova.network.neutron [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 730.212449] env[61906]: DEBUG nova.policy [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '83e30ca544ef4fd3b9fada2989066998', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd5257b698e5e4c359cb6e15c4fa548e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 730.472710] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d8a3b2-9b9d-4d6f-b561-0216f7fe01e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.481286] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861ea000-9827-4a2c-938d-0d198ace1cd4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.519123] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb8a41e-8fdb-489b-990c-70ef45426d49 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.525581] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68fb4191-c41b-4ffa-92f0-2495b2c99fd3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.541743] env[61906]: DEBUG nova.compute.provider_tree [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.640479] env[61906]: DEBUG nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 730.716437] env[61906]: DEBUG nova.network.neutron [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Successfully created port: c4140c45-ff57-41b0-b86f-24a25f5f62ba {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.778058] env[61906]: DEBUG nova.network.neutron [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.952681] env[61906]: DEBUG nova.network.neutron [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.046014] env[61906]: DEBUG nova.scheduler.client.report [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.288478] env[61906]: DEBUG nova.compute.manager [req-6ec3f171-b2b5-43db-afe7-c239c97f5b56 req-15b1b6a1-8197-4848-a87e-34512218599f service nova] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Received event network-vif-deleted-d4ce567a-fd81-426f-af76-0a6fe4e164c1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.457199] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Releasing lock "refresh_cache-83c86407-16b6-4025-9568-0c9c73d20b10" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.457607] env[61906]: DEBUG nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 731.457788] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 731.458188] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e754bb41-1a5b-4ba5-8140-c0a1e59f4f85 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.467407] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330b202c-f1bd-4cff-9afa-acf6a9c88d47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.493974] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 83c86407-16b6-4025-9568-0c9c73d20b10 could not be found. [ 731.493974] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 731.494166] env[61906]: INFO nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Took 0.04 seconds to destroy the instance on the hypervisor. [ 731.494437] env[61906]: DEBUG oslo.service.loopingcall [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.496849] env[61906]: DEBUG nova.compute.manager [-] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.496976] env[61906]: DEBUG nova.network.neutron [-] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.515575] env[61906]: DEBUG nova.network.neutron [-] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.550987] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.927s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.551670] env[61906]: ERROR nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7b51d2d4-73a2-4d32-a5df-c0a27a33336c, please check neutron logs for more information. [ 731.551670] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Traceback (most recent call last): [ 731.551670] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 731.551670] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] self.driver.spawn(context, instance, image_meta, [ 731.551670] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 731.551670] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] self._vmops.spawn(context, instance, image_meta, injected_files, [ 731.551670] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 731.551670] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] vm_ref = self.build_virtual_machine(instance, [ 731.551670] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 731.551670] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] vif_infos = vmwarevif.get_vif_info(self._session, [ 731.551670] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 731.552169] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] for vif in network_info: [ 731.552169] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 731.552169] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] return self._sync_wrapper(fn, *args, **kwargs) [ 731.552169] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 731.552169] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] self.wait() [ 731.552169] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 731.552169] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] self[:] = self._gt.wait() [ 731.552169] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 731.552169] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] return self._exit_event.wait() [ 731.552169] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 731.552169] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] result = hub.switch() [ 731.552169] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 731.552169] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] return self.greenlet.switch() [ 731.552992] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.552992] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] result = function(*args, **kwargs) [ 731.552992] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 731.552992] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] return func(*args, **kwargs) [ 731.552992] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.552992] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] raise e [ 731.552992] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.552992] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] nwinfo = self.network_api.allocate_for_instance( [ 731.552992] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.552992] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] created_port_ids = self._update_ports_for_instance( [ 731.552992] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.552992] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] with excutils.save_and_reraise_exception(): [ 731.552992] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.553489] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] self.force_reraise() [ 731.553489] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.553489] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] raise self.value [ 731.553489] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.553489] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] updated_port = self._update_port( [ 731.553489] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.553489] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] _ensure_no_port_binding_failure(port) [ 731.553489] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.553489] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] raise exception.PortBindingFailed(port_id=port['id']) [ 731.553489] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] nova.exception.PortBindingFailed: Binding failed for port 7b51d2d4-73a2-4d32-a5df-c0a27a33336c, please check neutron logs for more information. [ 731.553489] env[61906]: ERROR nova.compute.manager [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] [ 731.553968] env[61906]: DEBUG nova.compute.utils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Binding failed for port 7b51d2d4-73a2-4d32-a5df-c0a27a33336c, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 731.553968] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.729s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.557602] env[61906]: INFO nova.compute.claims [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.560475] env[61906]: DEBUG nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Build of instance 3c0602a1-ad7d-4f85-b4e0-d36565e28a38 was re-scheduled: Binding failed for port 7b51d2d4-73a2-4d32-a5df-c0a27a33336c, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 731.564112] env[61906]: DEBUG nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 731.564112] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Acquiring lock "refresh_cache-3c0602a1-ad7d-4f85-b4e0-d36565e28a38" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.564112] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Acquired lock "refresh_cache-3c0602a1-ad7d-4f85-b4e0-d36565e28a38" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.564112] env[61906]: DEBUG nova.network.neutron [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 731.651646] env[61906]: DEBUG nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 731.685583] env[61906]: DEBUG nova.virt.hardware [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 731.685583] env[61906]: DEBUG nova.virt.hardware [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 731.685583] env[61906]: DEBUG nova.virt.hardware [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.686029] env[61906]: DEBUG nova.virt.hardware [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 731.686029] env[61906]: DEBUG nova.virt.hardware [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.686208] env[61906]: DEBUG nova.virt.hardware [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 731.687102] env[61906]: DEBUG nova.virt.hardware [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 731.687102] env[61906]: DEBUG nova.virt.hardware [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 731.687102] env[61906]: DEBUG nova.virt.hardware [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 731.687102] env[61906]: DEBUG nova.virt.hardware [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 731.687102] env[61906]: DEBUG nova.virt.hardware [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 731.689546] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bade2819-fd7d-44c7-86fb-3076278663a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.698964] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc48507-9fe5-4e95-b414-81896c36e6cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.811650] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.811898] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.019675] env[61906]: DEBUG nova.network.neutron [-] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.082848] env[61906]: DEBUG nova.network.neutron [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.168500] env[61906]: DEBUG nova.network.neutron [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.266792] env[61906]: ERROR nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c4140c45-ff57-41b0-b86f-24a25f5f62ba, please check neutron logs for more information. [ 732.266792] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 732.266792] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.266792] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 732.266792] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.266792] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 732.266792] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.266792] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 732.266792] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.266792] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 732.266792] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.266792] env[61906]: ERROR nova.compute.manager raise self.value [ 732.266792] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.266792] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 732.266792] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.266792] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 732.267354] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.267354] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 732.267354] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c4140c45-ff57-41b0-b86f-24a25f5f62ba, please check neutron logs for more information. [ 732.267354] env[61906]: ERROR nova.compute.manager [ 732.267354] env[61906]: Traceback (most recent call last): [ 732.267354] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 732.267354] env[61906]: listener.cb(fileno) [ 732.267354] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.267354] env[61906]: result = function(*args, **kwargs) [ 732.267354] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.267354] env[61906]: return func(*args, **kwargs) [ 732.267354] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.267354] env[61906]: raise e [ 732.267354] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.267354] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 732.267354] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.267354] env[61906]: created_port_ids = self._update_ports_for_instance( [ 732.267354] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.267354] env[61906]: with excutils.save_and_reraise_exception(): [ 732.267354] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.267354] env[61906]: self.force_reraise() [ 732.267354] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.267354] env[61906]: raise self.value [ 732.267354] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.267354] env[61906]: updated_port = self._update_port( [ 732.267354] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.267354] env[61906]: _ensure_no_port_binding_failure(port) [ 732.267354] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.267354] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 732.268281] env[61906]: nova.exception.PortBindingFailed: Binding failed for port c4140c45-ff57-41b0-b86f-24a25f5f62ba, please check neutron logs for more information. [ 732.268281] env[61906]: Removing descriptor: 15 [ 732.268281] env[61906]: ERROR nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c4140c45-ff57-41b0-b86f-24a25f5f62ba, please check neutron logs for more information. [ 732.268281] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Traceback (most recent call last): [ 732.268281] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 732.268281] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] yield resources [ 732.268281] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.268281] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] self.driver.spawn(context, instance, image_meta, [ 732.268281] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 732.268281] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.268281] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.268281] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] vm_ref = self.build_virtual_machine(instance, [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] for vif in network_info: [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] return self._sync_wrapper(fn, *args, **kwargs) [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] self.wait() [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] self[:] = self._gt.wait() [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] return self._exit_event.wait() [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 732.268773] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] result = hub.switch() [ 732.269199] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 732.269199] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] return self.greenlet.switch() [ 732.269199] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.269199] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] result = function(*args, **kwargs) [ 732.269199] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.269199] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] return func(*args, **kwargs) [ 732.269199] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.269199] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] raise e [ 732.269199] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.269199] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] nwinfo = self.network_api.allocate_for_instance( [ 732.269199] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.269199] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] created_port_ids = self._update_ports_for_instance( [ 732.269199] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.269561] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] with excutils.save_and_reraise_exception(): [ 732.269561] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.269561] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] self.force_reraise() [ 732.269561] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.269561] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] raise self.value [ 732.269561] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.269561] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] updated_port = self._update_port( [ 732.269561] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.269561] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] _ensure_no_port_binding_failure(port) [ 732.269561] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.269561] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] raise exception.PortBindingFailed(port_id=port['id']) [ 732.269561] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] nova.exception.PortBindingFailed: Binding failed for port c4140c45-ff57-41b0-b86f-24a25f5f62ba, please check neutron logs for more information. [ 732.269561] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] [ 732.269933] env[61906]: INFO nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Terminating instance [ 732.272127] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Acquiring lock "refresh_cache-2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.272227] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Acquired lock "refresh_cache-2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.272360] env[61906]: DEBUG nova.network.neutron [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.523071] env[61906]: INFO nova.compute.manager [-] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Took 1.03 seconds to deallocate network for instance. [ 732.525107] env[61906]: DEBUG nova.compute.claims [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 732.525560] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.670754] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Releasing lock "refresh_cache-3c0602a1-ad7d-4f85-b4e0-d36565e28a38" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.670981] env[61906]: DEBUG nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 732.672080] env[61906]: DEBUG nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 732.672080] env[61906]: DEBUG nova.network.neutron [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.687811] env[61906]: DEBUG nova.network.neutron [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.796187] env[61906]: DEBUG nova.network.neutron [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.897885] env[61906]: DEBUG nova.network.neutron [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.910490] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90548fb6-c736-4e3b-b837-fe3007ca8522 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.919134] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82bd3be8-5c02-4109-9960-7d8284f7060d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.950529] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70369550-15a5-46a5-9022-b071b0fbe3fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.957896] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea5aa62-a817-48a1-a668-ecc3627dcdcc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.975109] env[61906]: DEBUG nova.compute.provider_tree [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.190791] env[61906]: DEBUG nova.network.neutron [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.353630] env[61906]: DEBUG nova.compute.manager [req-223fb32c-20b0-4765-9024-ed80db305a4f req-590957da-fc4a-4d8f-bc8d-828b00a5db6b service nova] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Received event network-changed-c4140c45-ff57-41b0-b86f-24a25f5f62ba {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.353832] env[61906]: DEBUG nova.compute.manager [req-223fb32c-20b0-4765-9024-ed80db305a4f req-590957da-fc4a-4d8f-bc8d-828b00a5db6b service nova] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Refreshing instance network info cache due to event network-changed-c4140c45-ff57-41b0-b86f-24a25f5f62ba. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 733.354063] env[61906]: DEBUG oslo_concurrency.lockutils [req-223fb32c-20b0-4765-9024-ed80db305a4f req-590957da-fc4a-4d8f-bc8d-828b00a5db6b service nova] Acquiring lock "refresh_cache-2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.402859] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Releasing lock "refresh_cache-2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.403471] env[61906]: DEBUG nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 733.403568] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.403853] env[61906]: DEBUG oslo_concurrency.lockutils [req-223fb32c-20b0-4765-9024-ed80db305a4f req-590957da-fc4a-4d8f-bc8d-828b00a5db6b service nova] Acquired lock "refresh_cache-2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.404089] env[61906]: DEBUG nova.network.neutron [req-223fb32c-20b0-4765-9024-ed80db305a4f req-590957da-fc4a-4d8f-bc8d-828b00a5db6b service nova] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Refreshing network info cache for port c4140c45-ff57-41b0-b86f-24a25f5f62ba {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 733.405170] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d84ee8db-ecb5-40c6-8858-e263720da429 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.414568] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4361f71b-2ca3-4a72-bf97-4504da318034 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.441737] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9 could not be found. [ 733.441973] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 733.442180] env[61906]: INFO nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 733.442422] env[61906]: DEBUG oslo.service.loopingcall [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.442675] env[61906]: DEBUG nova.compute.manager [-] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.442769] env[61906]: DEBUG nova.network.neutron [-] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.459586] env[61906]: DEBUG nova.network.neutron [-] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.479468] env[61906]: DEBUG nova.scheduler.client.report [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.672418] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "79452791-59cb-4722-bb4a-8e59d8c4e641" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.672694] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "79452791-59cb-4722-bb4a-8e59d8c4e641" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.696495] env[61906]: INFO nova.compute.manager [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] [instance: 3c0602a1-ad7d-4f85-b4e0-d36565e28a38] Took 1.02 seconds to deallocate network for instance. [ 733.925837] env[61906]: DEBUG nova.network.neutron [req-223fb32c-20b0-4765-9024-ed80db305a4f req-590957da-fc4a-4d8f-bc8d-828b00a5db6b service nova] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.961686] env[61906]: DEBUG nova.network.neutron [-] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.987738] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.988298] env[61906]: DEBUG nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.992045] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.922s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.992045] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.992045] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 733.992285] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.335s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.993682] env[61906]: INFO nova.compute.claims [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.998561] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98cb77c5-694e-43d9-a767-ca4b1b491fe1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.005536] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-750a5419-923e-4e8c-93f2-5cc22a6868bf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.010361] env[61906]: DEBUG nova.network.neutron [req-223fb32c-20b0-4765-9024-ed80db305a4f req-590957da-fc4a-4d8f-bc8d-828b00a5db6b service nova] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.023121] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de440d61-4198-44ea-ae3a-7081100cfeb4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.030573] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3530e0b8-8669-4f3a-b545-13e62848305d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.062036] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181503MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 734.062250] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.066387] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "7bae658e-b050-4639-b34e-c2671ef5c773" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.066603] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7bae658e-b050-4639-b34e-c2671ef5c773" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.467092] env[61906]: INFO nova.compute.manager [-] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Took 1.02 seconds to deallocate network for instance. [ 734.469543] env[61906]: DEBUG nova.compute.claims [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 734.469724] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.498300] env[61906]: DEBUG nova.compute.utils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 734.499951] env[61906]: DEBUG nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 734.500118] env[61906]: DEBUG nova.network.neutron [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 734.523578] env[61906]: DEBUG oslo_concurrency.lockutils [req-223fb32c-20b0-4765-9024-ed80db305a4f req-590957da-fc4a-4d8f-bc8d-828b00a5db6b service nova] Releasing lock "refresh_cache-2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.523858] env[61906]: DEBUG nova.compute.manager [req-223fb32c-20b0-4765-9024-ed80db305a4f req-590957da-fc4a-4d8f-bc8d-828b00a5db6b service nova] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Received event network-vif-deleted-c4140c45-ff57-41b0-b86f-24a25f5f62ba {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 734.556473] env[61906]: DEBUG nova.policy [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f752c49a92b14aecaa9c9e7a477ab2a0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1371191aa1c48e1a833735e0498b5bb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.730926] env[61906]: INFO nova.scheduler.client.report [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Deleted allocations for instance 3c0602a1-ad7d-4f85-b4e0-d36565e28a38 [ 734.900203] env[61906]: DEBUG nova.network.neutron [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Successfully created port: 455aa184-2c4a-450f-b1f4-4288b690df03 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 735.004232] env[61906]: DEBUG nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 735.245346] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2db1b201-1598-458b-bb00-f04d33cd5298 tempest-ServerRescueTestJSONUnderV235-70851400 tempest-ServerRescueTestJSONUnderV235-70851400-project-member] Lock "3c0602a1-ad7d-4f85-b4e0-d36565e28a38" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.993s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.298813] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b5fc18-b7a8-459a-a3db-486c0d5edfdb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.306266] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66804831-c071-46a7-bda6-fb4d8e0cb8c5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.336386] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a307be6-666f-46cd-923a-caa9471d798a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.343074] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11417765-4979-45db-a20a-0921d8b36744 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.355568] env[61906]: DEBUG nova.compute.provider_tree [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.591326] env[61906]: DEBUG nova.compute.manager [req-8f4245a0-24db-46cc-881a-c031e0ff8757 req-3a60b8fa-e135-4c81-9d0e-f980e4df552c service nova] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Received event network-changed-455aa184-2c4a-450f-b1f4-4288b690df03 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.591326] env[61906]: DEBUG nova.compute.manager [req-8f4245a0-24db-46cc-881a-c031e0ff8757 req-3a60b8fa-e135-4c81-9d0e-f980e4df552c service nova] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Refreshing instance network info cache due to event network-changed-455aa184-2c4a-450f-b1f4-4288b690df03. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 735.591326] env[61906]: DEBUG oslo_concurrency.lockutils [req-8f4245a0-24db-46cc-881a-c031e0ff8757 req-3a60b8fa-e135-4c81-9d0e-f980e4df552c service nova] Acquiring lock "refresh_cache-e0ea5eb3-82cf-44f1-91ee-2c2861aefee4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.591326] env[61906]: DEBUG oslo_concurrency.lockutils [req-8f4245a0-24db-46cc-881a-c031e0ff8757 req-3a60b8fa-e135-4c81-9d0e-f980e4df552c service nova] Acquired lock "refresh_cache-e0ea5eb3-82cf-44f1-91ee-2c2861aefee4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.591588] env[61906]: DEBUG nova.network.neutron [req-8f4245a0-24db-46cc-881a-c031e0ff8757 req-3a60b8fa-e135-4c81-9d0e-f980e4df552c service nova] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Refreshing network info cache for port 455aa184-2c4a-450f-b1f4-4288b690df03 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 735.749199] env[61906]: DEBUG nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 735.835255] env[61906]: ERROR nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 455aa184-2c4a-450f-b1f4-4288b690df03, please check neutron logs for more information. [ 735.835255] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 735.835255] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.835255] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 735.835255] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.835255] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 735.835255] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.835255] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 735.835255] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.835255] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 735.835255] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.835255] env[61906]: ERROR nova.compute.manager raise self.value [ 735.835255] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.835255] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 735.835255] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.835255] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 735.835763] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.835763] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 735.835763] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 455aa184-2c4a-450f-b1f4-4288b690df03, please check neutron logs for more information. [ 735.835763] env[61906]: ERROR nova.compute.manager [ 735.835763] env[61906]: Traceback (most recent call last): [ 735.835763] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 735.835763] env[61906]: listener.cb(fileno) [ 735.835763] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.835763] env[61906]: result = function(*args, **kwargs) [ 735.835763] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 735.835763] env[61906]: return func(*args, **kwargs) [ 735.835763] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.835763] env[61906]: raise e [ 735.835763] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.835763] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 735.835763] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.835763] env[61906]: created_port_ids = self._update_ports_for_instance( [ 735.835763] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.835763] env[61906]: with excutils.save_and_reraise_exception(): [ 735.835763] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.835763] env[61906]: self.force_reraise() [ 735.835763] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.835763] env[61906]: raise self.value [ 735.835763] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.835763] env[61906]: updated_port = self._update_port( [ 735.835763] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.835763] env[61906]: _ensure_no_port_binding_failure(port) [ 735.835763] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.835763] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 735.836586] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 455aa184-2c4a-450f-b1f4-4288b690df03, please check neutron logs for more information. [ 735.836586] env[61906]: Removing descriptor: 15 [ 735.858225] env[61906]: DEBUG nova.scheduler.client.report [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.017128] env[61906]: DEBUG nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 736.043997] env[61906]: DEBUG nova.virt.hardware [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 736.044311] env[61906]: DEBUG nova.virt.hardware [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 736.044472] env[61906]: DEBUG nova.virt.hardware [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.044651] env[61906]: DEBUG nova.virt.hardware [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 736.044793] env[61906]: DEBUG nova.virt.hardware [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.044940] env[61906]: DEBUG nova.virt.hardware [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 736.045184] env[61906]: DEBUG nova.virt.hardware [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 736.045344] env[61906]: DEBUG nova.virt.hardware [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 736.045525] env[61906]: DEBUG nova.virt.hardware [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 736.045724] env[61906]: DEBUG nova.virt.hardware [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 736.045905] env[61906]: DEBUG nova.virt.hardware [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.046791] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a371a9e-3e8c-4421-b21f-400ee3c5923e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.055445] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29acaf13-a5f5-476c-94ff-2ed1e00950da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.069219] env[61906]: ERROR nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 455aa184-2c4a-450f-b1f4-4288b690df03, please check neutron logs for more information. [ 736.069219] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Traceback (most recent call last): [ 736.069219] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 736.069219] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] yield resources [ 736.069219] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 736.069219] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] self.driver.spawn(context, instance, image_meta, [ 736.069219] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 736.069219] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 736.069219] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 736.069219] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] vm_ref = self.build_virtual_machine(instance, [ 736.069219] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 736.069613] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] vif_infos = vmwarevif.get_vif_info(self._session, [ 736.069613] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 736.069613] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] for vif in network_info: [ 736.069613] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 736.069613] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] return self._sync_wrapper(fn, *args, **kwargs) [ 736.069613] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 736.069613] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] self.wait() [ 736.069613] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 736.069613] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] self[:] = self._gt.wait() [ 736.069613] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 736.069613] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] return self._exit_event.wait() [ 736.069613] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 736.069613] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] current.throw(*self._exc) [ 736.070036] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.070036] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] result = function(*args, **kwargs) [ 736.070036] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 736.070036] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] return func(*args, **kwargs) [ 736.070036] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.070036] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] raise e [ 736.070036] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.070036] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] nwinfo = self.network_api.allocate_for_instance( [ 736.070036] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.070036] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] created_port_ids = self._update_ports_for_instance( [ 736.070036] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.070036] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] with excutils.save_and_reraise_exception(): [ 736.070036] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.070444] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] self.force_reraise() [ 736.070444] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.070444] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] raise self.value [ 736.070444] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.070444] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] updated_port = self._update_port( [ 736.070444] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.070444] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] _ensure_no_port_binding_failure(port) [ 736.070444] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.070444] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] raise exception.PortBindingFailed(port_id=port['id']) [ 736.070444] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] nova.exception.PortBindingFailed: Binding failed for port 455aa184-2c4a-450f-b1f4-4288b690df03, please check neutron logs for more information. [ 736.070444] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] [ 736.070444] env[61906]: INFO nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Terminating instance [ 736.073798] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Acquiring lock "refresh_cache-e0ea5eb3-82cf-44f1-91ee-2c2861aefee4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.111990] env[61906]: DEBUG nova.network.neutron [req-8f4245a0-24db-46cc-881a-c031e0ff8757 req-3a60b8fa-e135-4c81-9d0e-f980e4df552c service nova] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.197346] env[61906]: DEBUG nova.network.neutron [req-8f4245a0-24db-46cc-881a-c031e0ff8757 req-3a60b8fa-e135-4c81-9d0e-f980e4df552c service nova] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.270207] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.364736] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.365299] env[61906]: DEBUG nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 736.367784] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.566s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.699413] env[61906]: DEBUG oslo_concurrency.lockutils [req-8f4245a0-24db-46cc-881a-c031e0ff8757 req-3a60b8fa-e135-4c81-9d0e-f980e4df552c service nova] Releasing lock "refresh_cache-e0ea5eb3-82cf-44f1-91ee-2c2861aefee4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.700462] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Acquired lock "refresh_cache-e0ea5eb3-82cf-44f1-91ee-2c2861aefee4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.700813] env[61906]: DEBUG nova.network.neutron [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.875327] env[61906]: DEBUG nova.compute.utils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.877549] env[61906]: DEBUG nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 736.877707] env[61906]: DEBUG nova.network.neutron [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 736.923739] env[61906]: DEBUG nova.policy [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ec22db431334e7886acb497ebfc7eb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ac9a6686e7184fb1913e84a77985b449', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 737.229664] env[61906]: DEBUG nova.network.neutron [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.234240] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1af109a-3a4e-4643-9cb1-936558e8a93d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.241750] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe47a8cb-87b5-4be6-bff2-539a70b52f0d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.271458] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ed946f-8dac-410e-85ea-86b91ae0daaa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.279831] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ea62af-57d7-4a35-ae4b-a778a7660fad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.298036] env[61906]: DEBUG nova.compute.provider_tree [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.350959] env[61906]: DEBUG nova.network.neutron [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.381058] env[61906]: DEBUG nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 737.451833] env[61906]: DEBUG nova.network.neutron [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Successfully created port: be810f4e-b293-4b85-b0dc-133b33010e8a {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.620407] env[61906]: DEBUG nova.compute.manager [req-13147820-c09f-45f8-9aec-b6d7a41f82ef req-abe5ed78-ce9e-43fd-8883-f1919de7f9c8 service nova] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Received event network-vif-deleted-455aa184-2c4a-450f-b1f4-4288b690df03 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.800132] env[61906]: DEBUG nova.scheduler.client.report [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.853550] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Releasing lock "refresh_cache-e0ea5eb3-82cf-44f1-91ee-2c2861aefee4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.854050] env[61906]: DEBUG nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 737.854272] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 737.854570] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6034607c-3eec-420e-952e-c66e0e4fcf18 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.866537] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20a812d-ae55-437c-b702-80532f780566 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.892458] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e0ea5eb3-82cf-44f1-91ee-2c2861aefee4 could not be found. [ 737.892723] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 737.892846] env[61906]: INFO nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 737.893102] env[61906]: DEBUG oslo.service.loopingcall [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.893545] env[61906]: DEBUG nova.compute.manager [-] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.893638] env[61906]: DEBUG nova.network.neutron [-] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 737.912010] env[61906]: DEBUG nova.network.neutron [-] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.305441] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.937s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.305953] env[61906]: ERROR nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 909abf08-0182-4d09-a329-3d5a76441acb, please check neutron logs for more information. [ 738.305953] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Traceback (most recent call last): [ 738.305953] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.305953] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] self.driver.spawn(context, instance, image_meta, [ 738.305953] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 738.305953] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.305953] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.305953] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] vm_ref = self.build_virtual_machine(instance, [ 738.305953] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.305953] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.305953] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.306374] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] for vif in network_info: [ 738.306374] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.306374] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] return self._sync_wrapper(fn, *args, **kwargs) [ 738.306374] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.306374] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] self.wait() [ 738.306374] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.306374] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] self[:] = self._gt.wait() [ 738.306374] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.306374] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] return self._exit_event.wait() [ 738.306374] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 738.306374] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] result = hub.switch() [ 738.306374] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 738.306374] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] return self.greenlet.switch() [ 738.306864] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.306864] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] result = function(*args, **kwargs) [ 738.306864] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 738.306864] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] return func(*args, **kwargs) [ 738.306864] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.306864] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] raise e [ 738.306864] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.306864] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] nwinfo = self.network_api.allocate_for_instance( [ 738.306864] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.306864] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] created_port_ids = self._update_ports_for_instance( [ 738.306864] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.306864] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] with excutils.save_and_reraise_exception(): [ 738.306864] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.307328] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] self.force_reraise() [ 738.307328] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.307328] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] raise self.value [ 738.307328] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.307328] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] updated_port = self._update_port( [ 738.307328] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.307328] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] _ensure_no_port_binding_failure(port) [ 738.307328] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.307328] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] raise exception.PortBindingFailed(port_id=port['id']) [ 738.307328] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] nova.exception.PortBindingFailed: Binding failed for port 909abf08-0182-4d09-a329-3d5a76441acb, please check neutron logs for more information. [ 738.307328] env[61906]: ERROR nova.compute.manager [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] [ 738.307681] env[61906]: DEBUG nova.compute.utils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Binding failed for port 909abf08-0182-4d09-a329-3d5a76441acb, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 738.308074] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.975s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.313649] env[61906]: DEBUG nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Build of instance 4a2ed7bd-eb49-49a0-ba81-3c29baea6533 was re-scheduled: Binding failed for port 909abf08-0182-4d09-a329-3d5a76441acb, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 738.314163] env[61906]: DEBUG nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 738.315451] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquiring lock "refresh_cache-4a2ed7bd-eb49-49a0-ba81-3c29baea6533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.315451] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquired lock "refresh_cache-4a2ed7bd-eb49-49a0-ba81-3c29baea6533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.315451] env[61906]: DEBUG nova.network.neutron [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.395582] env[61906]: DEBUG nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 738.414118] env[61906]: DEBUG nova.network.neutron [-] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.428206] env[61906]: DEBUG nova.virt.hardware [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 738.428464] env[61906]: DEBUG nova.virt.hardware [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 738.428617] env[61906]: DEBUG nova.virt.hardware [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.428796] env[61906]: DEBUG nova.virt.hardware [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 738.428942] env[61906]: DEBUG nova.virt.hardware [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.429105] env[61906]: DEBUG nova.virt.hardware [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 738.429361] env[61906]: DEBUG nova.virt.hardware [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 738.429523] env[61906]: DEBUG nova.virt.hardware [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 738.429689] env[61906]: DEBUG nova.virt.hardware [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 738.429848] env[61906]: DEBUG nova.virt.hardware [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 738.430227] env[61906]: DEBUG nova.virt.hardware [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 738.431097] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40331ce-4dbf-4530-a309-994505e8786e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.440728] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8a2603-f99d-4384-a520-7d743f8a0f07 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.524335] env[61906]: ERROR nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port be810f4e-b293-4b85-b0dc-133b33010e8a, please check neutron logs for more information. [ 738.524335] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 738.524335] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.524335] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 738.524335] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.524335] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 738.524335] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.524335] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 738.524335] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.524335] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 738.524335] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.524335] env[61906]: ERROR nova.compute.manager raise self.value [ 738.524335] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.524335] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 738.524335] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.524335] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 738.525143] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.525143] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 738.525143] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port be810f4e-b293-4b85-b0dc-133b33010e8a, please check neutron logs for more information. [ 738.525143] env[61906]: ERROR nova.compute.manager [ 738.525143] env[61906]: Traceback (most recent call last): [ 738.525143] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 738.525143] env[61906]: listener.cb(fileno) [ 738.525143] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.525143] env[61906]: result = function(*args, **kwargs) [ 738.525143] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 738.525143] env[61906]: return func(*args, **kwargs) [ 738.525143] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.525143] env[61906]: raise e [ 738.525143] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.525143] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 738.525143] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.525143] env[61906]: created_port_ids = self._update_ports_for_instance( [ 738.525143] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.525143] env[61906]: with excutils.save_and_reraise_exception(): [ 738.525143] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.525143] env[61906]: self.force_reraise() [ 738.525143] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.525143] env[61906]: raise self.value [ 738.525143] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.525143] env[61906]: updated_port = self._update_port( [ 738.525143] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.525143] env[61906]: _ensure_no_port_binding_failure(port) [ 738.525143] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.525143] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 738.526203] env[61906]: nova.exception.PortBindingFailed: Binding failed for port be810f4e-b293-4b85-b0dc-133b33010e8a, please check neutron logs for more information. [ 738.526203] env[61906]: Removing descriptor: 15 [ 738.526203] env[61906]: ERROR nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port be810f4e-b293-4b85-b0dc-133b33010e8a, please check neutron logs for more information. [ 738.526203] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Traceback (most recent call last): [ 738.526203] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 738.526203] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] yield resources [ 738.526203] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.526203] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] self.driver.spawn(context, instance, image_meta, [ 738.526203] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 738.526203] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.526203] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.526203] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] vm_ref = self.build_virtual_machine(instance, [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] for vif in network_info: [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] return self._sync_wrapper(fn, *args, **kwargs) [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] self.wait() [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] self[:] = self._gt.wait() [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] return self._exit_event.wait() [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 738.526730] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] result = hub.switch() [ 738.527235] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 738.527235] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] return self.greenlet.switch() [ 738.527235] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.527235] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] result = function(*args, **kwargs) [ 738.527235] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 738.527235] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] return func(*args, **kwargs) [ 738.527235] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.527235] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] raise e [ 738.527235] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.527235] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] nwinfo = self.network_api.allocate_for_instance( [ 738.527235] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.527235] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] created_port_ids = self._update_ports_for_instance( [ 738.527235] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.527716] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] with excutils.save_and_reraise_exception(): [ 738.527716] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.527716] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] self.force_reraise() [ 738.527716] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.527716] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] raise self.value [ 738.527716] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.527716] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] updated_port = self._update_port( [ 738.527716] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.527716] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] _ensure_no_port_binding_failure(port) [ 738.527716] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.527716] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] raise exception.PortBindingFailed(port_id=port['id']) [ 738.527716] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] nova.exception.PortBindingFailed: Binding failed for port be810f4e-b293-4b85-b0dc-133b33010e8a, please check neutron logs for more information. [ 738.527716] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] [ 738.528176] env[61906]: INFO nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Terminating instance [ 738.528176] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "refresh_cache-ce3c082c-a3cb-452d-9581-0bed6aea892b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.528176] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "refresh_cache-ce3c082c-a3cb-452d-9581-0bed6aea892b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.528176] env[61906]: DEBUG nova.network.neutron [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.834453] env[61906]: DEBUG nova.network.neutron [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.885807] env[61906]: DEBUG nova.network.neutron [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.917113] env[61906]: INFO nova.compute.manager [-] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Took 1.02 seconds to deallocate network for instance. [ 738.919252] env[61906]: DEBUG nova.compute.claims [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 738.920029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.053744] env[61906]: DEBUG nova.network.neutron [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.092908] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-826ddcde-f1a8-4ebc-9bfb-445aca53ae34 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.102688] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6132cbc1-087b-4f8f-aa47-495789aa430c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.132885] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0399db2-3549-478c-8e52-4d6e454066c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.141349] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81531a3-a43b-4986-99d0-b69e8566ef35 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.153977] env[61906]: DEBUG nova.compute.provider_tree [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.155990] env[61906]: DEBUG nova.network.neutron [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.388463] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Releasing lock "refresh_cache-4a2ed7bd-eb49-49a0-ba81-3c29baea6533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.388745] env[61906]: DEBUG nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 739.388924] env[61906]: DEBUG nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.389124] env[61906]: DEBUG nova.network.neutron [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.402751] env[61906]: DEBUG nova.network.neutron [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.642539] env[61906]: DEBUG nova.compute.manager [req-b69d71a2-b582-4714-a165-78173e045d87 req-a5a79312-e7fd-4e69-ad1c-d5caa5767973 service nova] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Received event network-changed-be810f4e-b293-4b85-b0dc-133b33010e8a {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.642737] env[61906]: DEBUG nova.compute.manager [req-b69d71a2-b582-4714-a165-78173e045d87 req-a5a79312-e7fd-4e69-ad1c-d5caa5767973 service nova] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Refreshing instance network info cache due to event network-changed-be810f4e-b293-4b85-b0dc-133b33010e8a. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 739.642929] env[61906]: DEBUG oslo_concurrency.lockutils [req-b69d71a2-b582-4714-a165-78173e045d87 req-a5a79312-e7fd-4e69-ad1c-d5caa5767973 service nova] Acquiring lock "refresh_cache-ce3c082c-a3cb-452d-9581-0bed6aea892b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.661689] env[61906]: DEBUG nova.scheduler.client.report [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.663021] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "refresh_cache-ce3c082c-a3cb-452d-9581-0bed6aea892b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.663223] env[61906]: DEBUG nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 739.663415] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 739.664052] env[61906]: DEBUG oslo_concurrency.lockutils [req-b69d71a2-b582-4714-a165-78173e045d87 req-a5a79312-e7fd-4e69-ad1c-d5caa5767973 service nova] Acquired lock "refresh_cache-ce3c082c-a3cb-452d-9581-0bed6aea892b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.664217] env[61906]: DEBUG nova.network.neutron [req-b69d71a2-b582-4714-a165-78173e045d87 req-a5a79312-e7fd-4e69-ad1c-d5caa5767973 service nova] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Refreshing network info cache for port be810f4e-b293-4b85-b0dc-133b33010e8a {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 739.665124] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1740c09c-ed1c-426d-a4ce-f70c2949f558 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.675035] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed086ca2-47fe-4cb5-9857-b91c85f62e13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.696853] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ce3c082c-a3cb-452d-9581-0bed6aea892b could not be found. [ 739.697084] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 739.697274] env[61906]: INFO nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 739.697515] env[61906]: DEBUG oslo.service.loopingcall [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.697733] env[61906]: DEBUG nova.compute.manager [-] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.697846] env[61906]: DEBUG nova.network.neutron [-] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.712362] env[61906]: DEBUG nova.network.neutron [-] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.905681] env[61906]: DEBUG nova.network.neutron [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.169017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.861s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.170094] env[61906]: ERROR nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8, please check neutron logs for more information. [ 740.170094] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Traceback (most recent call last): [ 740.170094] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 740.170094] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] self.driver.spawn(context, instance, image_meta, [ 740.170094] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 740.170094] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.170094] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.170094] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] vm_ref = self.build_virtual_machine(instance, [ 740.170094] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.170094] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.170094] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.170544] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] for vif in network_info: [ 740.170544] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 740.170544] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] return self._sync_wrapper(fn, *args, **kwargs) [ 740.170544] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 740.170544] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] self.wait() [ 740.170544] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 740.170544] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] self[:] = self._gt.wait() [ 740.170544] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.170544] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] return self._exit_event.wait() [ 740.170544] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 740.170544] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] result = hub.switch() [ 740.170544] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 740.170544] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] return self.greenlet.switch() [ 740.170938] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.170938] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] result = function(*args, **kwargs) [ 740.170938] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 740.170938] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] return func(*args, **kwargs) [ 740.170938] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.170938] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] raise e [ 740.170938] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.170938] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] nwinfo = self.network_api.allocate_for_instance( [ 740.170938] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.170938] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] created_port_ids = self._update_ports_for_instance( [ 740.170938] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.170938] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] with excutils.save_and_reraise_exception(): [ 740.170938] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.171360] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] self.force_reraise() [ 740.171360] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.171360] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] raise self.value [ 740.171360] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.171360] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] updated_port = self._update_port( [ 740.171360] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.171360] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] _ensure_no_port_binding_failure(port) [ 740.171360] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.171360] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] raise exception.PortBindingFailed(port_id=port['id']) [ 740.171360] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] nova.exception.PortBindingFailed: Binding failed for port 3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8, please check neutron logs for more information. [ 740.171360] env[61906]: ERROR nova.compute.manager [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] [ 740.171713] env[61906]: DEBUG nova.compute.utils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Binding failed for port 3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 740.174034] env[61906]: DEBUG nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Build of instance 0dca3209-d37b-4536-bbf9-72f0418cef8e was re-scheduled: Binding failed for port 3d0cfa50-bb2a-4255-a1d3-03f2b2c76cb8, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 740.174473] env[61906]: DEBUG nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 740.174699] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquiring lock "refresh_cache-0dca3209-d37b-4536-bbf9-72f0418cef8e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.174844] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Acquired lock "refresh_cache-0dca3209-d37b-4536-bbf9-72f0418cef8e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.175167] env[61906]: DEBUG nova.network.neutron [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.176148] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.716s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.192932] env[61906]: DEBUG nova.network.neutron [req-b69d71a2-b582-4714-a165-78173e045d87 req-a5a79312-e7fd-4e69-ad1c-d5caa5767973 service nova] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.214711] env[61906]: DEBUG nova.network.neutron [-] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.302358] env[61906]: DEBUG nova.network.neutron [req-b69d71a2-b582-4714-a165-78173e045d87 req-a5a79312-e7fd-4e69-ad1c-d5caa5767973 service nova] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.408780] env[61906]: INFO nova.compute.manager [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 4a2ed7bd-eb49-49a0-ba81-3c29baea6533] Took 1.02 seconds to deallocate network for instance. [ 740.698478] env[61906]: DEBUG nova.network.neutron [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.717344] env[61906]: INFO nova.compute.manager [-] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Took 1.02 seconds to deallocate network for instance. [ 740.719854] env[61906]: DEBUG nova.compute.claims [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 740.720030] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.758608] env[61906]: DEBUG nova.network.neutron [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.807687] env[61906]: DEBUG oslo_concurrency.lockutils [req-b69d71a2-b582-4714-a165-78173e045d87 req-a5a79312-e7fd-4e69-ad1c-d5caa5767973 service nova] Releasing lock "refresh_cache-ce3c082c-a3cb-452d-9581-0bed6aea892b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.807943] env[61906]: DEBUG nova.compute.manager [req-b69d71a2-b582-4714-a165-78173e045d87 req-a5a79312-e7fd-4e69-ad1c-d5caa5767973 service nova] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Received event network-vif-deleted-be810f4e-b293-4b85-b0dc-133b33010e8a {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.993241] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33ccd4d6-ec4e-4c19-81c1-931dc26cedd3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.001411] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3820152-bacb-40ab-b1b1-37ddb39de119 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.031659] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49deda61-b16a-420b-8255-26006288b71e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.039526] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8942695b-ec6f-4b99-8fe3-9912344dc3da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.054366] env[61906]: DEBUG nova.compute.provider_tree [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.260719] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Releasing lock "refresh_cache-0dca3209-d37b-4536-bbf9-72f0418cef8e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.261038] env[61906]: DEBUG nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 741.261264] env[61906]: DEBUG nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.261327] env[61906]: DEBUG nova.network.neutron [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.276773] env[61906]: DEBUG nova.network.neutron [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.438745] env[61906]: INFO nova.scheduler.client.report [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Deleted allocations for instance 4a2ed7bd-eb49-49a0-ba81-3c29baea6533 [ 741.558089] env[61906]: DEBUG nova.scheduler.client.report [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.779549] env[61906]: DEBUG nova.network.neutron [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.950311] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fbaaafe-97ba-454a-8caa-5bfafe21978c tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Lock "4a2ed7bd-eb49-49a0-ba81-3c29baea6533" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.804s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.065047] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.888s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.065047] env[61906]: ERROR nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 95c18951-8f02-411b-9709-3412b832e328, please check neutron logs for more information. [ 742.065047] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Traceback (most recent call last): [ 742.065047] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 742.065047] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] self.driver.spawn(context, instance, image_meta, [ 742.065047] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 742.065047] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.065047] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.065047] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] vm_ref = self.build_virtual_machine(instance, [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] for vif in network_info: [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] return self._sync_wrapper(fn, *args, **kwargs) [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] self.wait() [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] self[:] = self._gt.wait() [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] return self._exit_event.wait() [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 742.065581] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] result = hub.switch() [ 742.065973] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 742.065973] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] return self.greenlet.switch() [ 742.065973] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.065973] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] result = function(*args, **kwargs) [ 742.065973] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 742.065973] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] return func(*args, **kwargs) [ 742.065973] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 742.065973] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] raise e [ 742.065973] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.065973] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] nwinfo = self.network_api.allocate_for_instance( [ 742.065973] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.065973] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] created_port_ids = self._update_ports_for_instance( [ 742.065973] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.066459] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] with excutils.save_and_reraise_exception(): [ 742.066459] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.066459] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] self.force_reraise() [ 742.066459] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.066459] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] raise self.value [ 742.066459] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.066459] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] updated_port = self._update_port( [ 742.066459] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.066459] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] _ensure_no_port_binding_failure(port) [ 742.066459] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.066459] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] raise exception.PortBindingFailed(port_id=port['id']) [ 742.066459] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] nova.exception.PortBindingFailed: Binding failed for port 95c18951-8f02-411b-9709-3412b832e328, please check neutron logs for more information. [ 742.066459] env[61906]: ERROR nova.compute.manager [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] [ 742.066850] env[61906]: DEBUG nova.compute.utils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Binding failed for port 95c18951-8f02-411b-9709-3412b832e328, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 742.066850] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.407s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.068073] env[61906]: INFO nova.compute.claims [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.070539] env[61906]: DEBUG nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Build of instance b1ec5d55-96a6-42eb-ba54-c4a3904e22cd was re-scheduled: Binding failed for port 95c18951-8f02-411b-9709-3412b832e328, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 742.070959] env[61906]: DEBUG nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 742.071204] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Acquiring lock "refresh_cache-b1ec5d55-96a6-42eb-ba54-c4a3904e22cd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.071353] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Acquired lock "refresh_cache-b1ec5d55-96a6-42eb-ba54-c4a3904e22cd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.071510] env[61906]: DEBUG nova.network.neutron [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.281736] env[61906]: INFO nova.compute.manager [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] [instance: 0dca3209-d37b-4536-bbf9-72f0418cef8e] Took 1.02 seconds to deallocate network for instance. [ 742.451778] env[61906]: DEBUG nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 742.592228] env[61906]: DEBUG nova.network.neutron [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.662725] env[61906]: DEBUG nova.network.neutron [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.975944] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.165537] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Releasing lock "refresh_cache-b1ec5d55-96a6-42eb-ba54-c4a3904e22cd" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.165707] env[61906]: DEBUG nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 743.165874] env[61906]: DEBUG nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.166058] env[61906]: DEBUG nova.network.neutron [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.185072] env[61906]: DEBUG nova.network.neutron [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.310174] env[61906]: INFO nova.scheduler.client.report [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Deleted allocations for instance 0dca3209-d37b-4536-bbf9-72f0418cef8e [ 743.335901] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c1820c-dc52-4cc8-993c-2463bdd9740e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.344458] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2d3ee0-429e-48a9-80c5-f1a283368233 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.376499] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e433e9ad-73c4-41b0-b071-1afba10e49c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.383710] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343d5b64-3a55-4400-aabd-bc0fc8380180 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.396710] env[61906]: DEBUG nova.compute.provider_tree [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.687655] env[61906]: DEBUG nova.network.neutron [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.818193] env[61906]: DEBUG oslo_concurrency.lockutils [None req-71e866e4-8c73-41fa-a7a5-2b5cc4d8f064 tempest-ListImageFiltersTestJSON-467127626 tempest-ListImageFiltersTestJSON-467127626-project-member] Lock "0dca3209-d37b-4536-bbf9-72f0418cef8e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.515s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.899724] env[61906]: DEBUG nova.scheduler.client.report [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.194293] env[61906]: INFO nova.compute.manager [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] [instance: b1ec5d55-96a6-42eb-ba54-c4a3904e22cd] Took 1.03 seconds to deallocate network for instance. [ 744.320989] env[61906]: DEBUG nova.compute.manager [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 744.405727] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.406253] env[61906]: DEBUG nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 744.408770] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.832s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.410239] env[61906]: INFO nova.compute.claims [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.842759] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.915030] env[61906]: DEBUG nova.compute.utils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.918567] env[61906]: DEBUG nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 744.918567] env[61906]: DEBUG nova.network.neutron [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 744.967246] env[61906]: DEBUG nova.policy [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f502ef6a6a6e42c2b28b04808bdb4fed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfb544eadaa348d1b61205a43a73cb65', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.248446] env[61906]: INFO nova.scheduler.client.report [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Deleted allocations for instance b1ec5d55-96a6-42eb-ba54-c4a3904e22cd [ 745.274899] env[61906]: DEBUG nova.network.neutron [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Successfully created port: 04d25abb-d1a7-496f-9ef8-6dfc268ee592 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.419880] env[61906]: DEBUG nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 745.742075] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13101e99-2524-424d-9e8e-8313018a45f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.750489] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2706979-dfe2-4085-9305-596392bde90a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.756662] env[61906]: DEBUG oslo_concurrency.lockutils [None req-492057c5-004c-4465-ad47-3b7ccaab6916 tempest-TenantUsagesTestJSON-1532269427 tempest-TenantUsagesTestJSON-1532269427-project-member] Lock "b1ec5d55-96a6-42eb-ba54-c4a3904e22cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.869s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.783802] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16afc72e-76dc-4a71-b49a-aa523c2104fd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.791511] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bac4a5d-fb48-445e-825b-d8c991cc3cac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.806759] env[61906]: DEBUG nova.compute.provider_tree [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.149951] env[61906]: DEBUG nova.compute.manager [req-1fdf8011-b2c9-45e8-9204-24d97e4170ad req-434fe716-ab9d-4407-aba9-ae947466328b service nova] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Received event network-changed-04d25abb-d1a7-496f-9ef8-6dfc268ee592 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.149951] env[61906]: DEBUG nova.compute.manager [req-1fdf8011-b2c9-45e8-9204-24d97e4170ad req-434fe716-ab9d-4407-aba9-ae947466328b service nova] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Refreshing instance network info cache due to event network-changed-04d25abb-d1a7-496f-9ef8-6dfc268ee592. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 746.150216] env[61906]: DEBUG oslo_concurrency.lockutils [req-1fdf8011-b2c9-45e8-9204-24d97e4170ad req-434fe716-ab9d-4407-aba9-ae947466328b service nova] Acquiring lock "refresh_cache-a8890726-6c64-4e09-a9a4-0726aa2e8c31" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.150394] env[61906]: DEBUG oslo_concurrency.lockutils [req-1fdf8011-b2c9-45e8-9204-24d97e4170ad req-434fe716-ab9d-4407-aba9-ae947466328b service nova] Acquired lock "refresh_cache-a8890726-6c64-4e09-a9a4-0726aa2e8c31" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.150688] env[61906]: DEBUG nova.network.neutron [req-1fdf8011-b2c9-45e8-9204-24d97e4170ad req-434fe716-ab9d-4407-aba9-ae947466328b service nova] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Refreshing network info cache for port 04d25abb-d1a7-496f-9ef8-6dfc268ee592 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 746.287291] env[61906]: DEBUG nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 746.313327] env[61906]: DEBUG nova.scheduler.client.report [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.399130] env[61906]: ERROR nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 04d25abb-d1a7-496f-9ef8-6dfc268ee592, please check neutron logs for more information. [ 746.399130] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 746.399130] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.399130] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 746.399130] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.399130] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 746.399130] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.399130] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 746.399130] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.399130] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 746.399130] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.399130] env[61906]: ERROR nova.compute.manager raise self.value [ 746.399130] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.399130] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 746.399130] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.399130] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 746.399693] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.399693] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 746.399693] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 04d25abb-d1a7-496f-9ef8-6dfc268ee592, please check neutron logs for more information. [ 746.399693] env[61906]: ERROR nova.compute.manager [ 746.399693] env[61906]: Traceback (most recent call last): [ 746.399693] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 746.399693] env[61906]: listener.cb(fileno) [ 746.399693] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.399693] env[61906]: result = function(*args, **kwargs) [ 746.399693] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 746.399693] env[61906]: return func(*args, **kwargs) [ 746.399693] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.399693] env[61906]: raise e [ 746.399693] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.399693] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 746.399693] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.399693] env[61906]: created_port_ids = self._update_ports_for_instance( [ 746.399693] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.399693] env[61906]: with excutils.save_and_reraise_exception(): [ 746.399693] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.399693] env[61906]: self.force_reraise() [ 746.399693] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.399693] env[61906]: raise self.value [ 746.399693] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.399693] env[61906]: updated_port = self._update_port( [ 746.399693] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.399693] env[61906]: _ensure_no_port_binding_failure(port) [ 746.399693] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.399693] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 746.400716] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 04d25abb-d1a7-496f-9ef8-6dfc268ee592, please check neutron logs for more information. [ 746.400716] env[61906]: Removing descriptor: 15 [ 746.437219] env[61906]: DEBUG nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 746.466789] env[61906]: DEBUG nova.virt.hardware [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.467306] env[61906]: DEBUG nova.virt.hardware [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.467306] env[61906]: DEBUG nova.virt.hardware [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.467494] env[61906]: DEBUG nova.virt.hardware [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.467535] env[61906]: DEBUG nova.virt.hardware [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.467702] env[61906]: DEBUG nova.virt.hardware [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.471272] env[61906]: DEBUG nova.virt.hardware [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.471272] env[61906]: DEBUG nova.virt.hardware [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.471272] env[61906]: DEBUG nova.virt.hardware [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.471272] env[61906]: DEBUG nova.virt.hardware [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.471272] env[61906]: DEBUG nova.virt.hardware [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.471515] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab3380b-7d6f-4b21-a3f8-7abf84a7869c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.480030] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85671265-7cd6-4dd6-a2c5-80be3aa66d66 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.497700] env[61906]: ERROR nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 04d25abb-d1a7-496f-9ef8-6dfc268ee592, please check neutron logs for more information. [ 746.497700] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Traceback (most recent call last): [ 746.497700] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 746.497700] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] yield resources [ 746.497700] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 746.497700] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] self.driver.spawn(context, instance, image_meta, [ 746.497700] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 746.497700] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.497700] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.497700] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] vm_ref = self.build_virtual_machine(instance, [ 746.497700] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.498142] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.498142] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.498142] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] for vif in network_info: [ 746.498142] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 746.498142] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] return self._sync_wrapper(fn, *args, **kwargs) [ 746.498142] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 746.498142] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] self.wait() [ 746.498142] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 746.498142] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] self[:] = self._gt.wait() [ 746.498142] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.498142] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] return self._exit_event.wait() [ 746.498142] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 746.498142] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] current.throw(*self._exc) [ 746.498534] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.498534] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] result = function(*args, **kwargs) [ 746.498534] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 746.498534] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] return func(*args, **kwargs) [ 746.498534] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.498534] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] raise e [ 746.498534] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.498534] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] nwinfo = self.network_api.allocate_for_instance( [ 746.498534] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.498534] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] created_port_ids = self._update_ports_for_instance( [ 746.498534] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.498534] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] with excutils.save_and_reraise_exception(): [ 746.498534] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.499010] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] self.force_reraise() [ 746.499010] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.499010] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] raise self.value [ 746.499010] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.499010] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] updated_port = self._update_port( [ 746.499010] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.499010] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] _ensure_no_port_binding_failure(port) [ 746.499010] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.499010] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] raise exception.PortBindingFailed(port_id=port['id']) [ 746.499010] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] nova.exception.PortBindingFailed: Binding failed for port 04d25abb-d1a7-496f-9ef8-6dfc268ee592, please check neutron logs for more information. [ 746.499010] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] [ 746.499010] env[61906]: INFO nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Terminating instance [ 746.499931] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Acquiring lock "refresh_cache-a8890726-6c64-4e09-a9a4-0726aa2e8c31" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.671776] env[61906]: DEBUG nova.network.neutron [req-1fdf8011-b2c9-45e8-9204-24d97e4170ad req-434fe716-ab9d-4407-aba9-ae947466328b service nova] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.748867] env[61906]: DEBUG nova.network.neutron [req-1fdf8011-b2c9-45e8-9204-24d97e4170ad req-434fe716-ab9d-4407-aba9-ae947466328b service nova] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.807626] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.821547] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.822066] env[61906]: DEBUG nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 746.824536] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.299s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.252470] env[61906]: DEBUG oslo_concurrency.lockutils [req-1fdf8011-b2c9-45e8-9204-24d97e4170ad req-434fe716-ab9d-4407-aba9-ae947466328b service nova] Releasing lock "refresh_cache-a8890726-6c64-4e09-a9a4-0726aa2e8c31" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.252879] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Acquired lock "refresh_cache-a8890726-6c64-4e09-a9a4-0726aa2e8c31" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.253166] env[61906]: DEBUG nova.network.neutron [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 747.331229] env[61906]: DEBUG nova.compute.utils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 747.337027] env[61906]: DEBUG nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 747.337027] env[61906]: DEBUG nova.network.neutron [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 747.408622] env[61906]: DEBUG nova.policy [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c0ff951d644f48ddb13606453633b7cf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b31840e9dcfd4d9abef70f6d8a44f3f6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 747.658025] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65872af-8fea-41eb-aee6-5008b4e048bf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.666950] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8c3d36-76fd-4a03-b1e3-cdd730a009a3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.711329] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992f8d94-7b16-403e-b5c2-84e21054468b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.718274] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66861420-4f65-473b-baba-39bf2d058b71 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.732536] env[61906]: DEBUG nova.compute.provider_tree [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.753353] env[61906]: DEBUG nova.network.neutron [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Successfully created port: cd611107-eb27-4ac8-b1f6-9a29c1516ac7 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 747.795780] env[61906]: DEBUG nova.network.neutron [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.837016] env[61906]: DEBUG nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 747.984281] env[61906]: DEBUG nova.network.neutron [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.179528] env[61906]: DEBUG nova.compute.manager [req-beb754af-ee89-4581-8db5-f88c160448d1 req-53f35e7f-5a88-4d80-92d8-940b67e06ea5 service nova] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Received event network-vif-deleted-04d25abb-d1a7-496f-9ef8-6dfc268ee592 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.239236] env[61906]: DEBUG nova.scheduler.client.report [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.487642] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Releasing lock "refresh_cache-a8890726-6c64-4e09-a9a4-0726aa2e8c31" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.488069] env[61906]: DEBUG nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 748.488269] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 748.488628] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e4155a1-ab85-4392-a61c-2e61da1a4279 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.498529] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b395bcb9-1957-43b2-a774-38b065c3b949 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.523370] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a8890726-6c64-4e09-a9a4-0726aa2e8c31 could not be found. [ 748.523499] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.523697] env[61906]: INFO nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Took 0.04 seconds to destroy the instance on the hypervisor. [ 748.524026] env[61906]: DEBUG oslo.service.loopingcall [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.524207] env[61906]: DEBUG nova.compute.manager [-] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.524302] env[61906]: DEBUG nova.network.neutron [-] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.551977] env[61906]: DEBUG nova.network.neutron [-] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.744255] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.920s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.744912] env[61906]: ERROR nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d4ce567a-fd81-426f-af76-0a6fe4e164c1, please check neutron logs for more information. [ 748.744912] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Traceback (most recent call last): [ 748.744912] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 748.744912] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] self.driver.spawn(context, instance, image_meta, [ 748.744912] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 748.744912] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] self._vmops.spawn(context, instance, image_meta, injected_files, [ 748.744912] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 748.744912] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] vm_ref = self.build_virtual_machine(instance, [ 748.744912] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 748.744912] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] vif_infos = vmwarevif.get_vif_info(self._session, [ 748.744912] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] for vif in network_info: [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] return self._sync_wrapper(fn, *args, **kwargs) [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] self.wait() [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] self[:] = self._gt.wait() [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] return self._exit_event.wait() [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] current.throw(*self._exc) [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] result = function(*args, **kwargs) [ 748.745311] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 748.745875] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] return func(*args, **kwargs) [ 748.745875] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 748.745875] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] raise e [ 748.745875] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 748.745875] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] nwinfo = self.network_api.allocate_for_instance( [ 748.745875] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 748.745875] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] created_port_ids = self._update_ports_for_instance( [ 748.745875] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 748.745875] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] with excutils.save_and_reraise_exception(): [ 748.745875] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 748.745875] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] self.force_reraise() [ 748.745875] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 748.745875] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] raise self.value [ 748.746324] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 748.746324] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] updated_port = self._update_port( [ 748.746324] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 748.746324] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] _ensure_no_port_binding_failure(port) [ 748.746324] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 748.746324] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] raise exception.PortBindingFailed(port_id=port['id']) [ 748.746324] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] nova.exception.PortBindingFailed: Binding failed for port d4ce567a-fd81-426f-af76-0a6fe4e164c1, please check neutron logs for more information. [ 748.746324] env[61906]: ERROR nova.compute.manager [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] [ 748.746324] env[61906]: DEBUG nova.compute.utils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Binding failed for port d4ce567a-fd81-426f-af76-0a6fe4e164c1, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 748.747055] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.685s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.749679] env[61906]: DEBUG nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Build of instance 83c86407-16b6-4025-9568-0c9c73d20b10 was re-scheduled: Binding failed for port d4ce567a-fd81-426f-af76-0a6fe4e164c1, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 748.750109] env[61906]: DEBUG nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 748.750340] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Acquiring lock "refresh_cache-83c86407-16b6-4025-9568-0c9c73d20b10" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.750484] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Acquired lock "refresh_cache-83c86407-16b6-4025-9568-0c9c73d20b10" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.750636] env[61906]: DEBUG nova.network.neutron [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 748.851484] env[61906]: DEBUG nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 748.884440] env[61906]: DEBUG nova.virt.hardware [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 748.884707] env[61906]: DEBUG nova.virt.hardware [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 748.884862] env[61906]: DEBUG nova.virt.hardware [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 748.885057] env[61906]: DEBUG nova.virt.hardware [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 748.885278] env[61906]: DEBUG nova.virt.hardware [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 748.885483] env[61906]: DEBUG nova.virt.hardware [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 748.885653] env[61906]: DEBUG nova.virt.hardware [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 748.885809] env[61906]: DEBUG nova.virt.hardware [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 748.885972] env[61906]: DEBUG nova.virt.hardware [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 748.886929] env[61906]: DEBUG nova.virt.hardware [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 748.887083] env[61906]: DEBUG nova.virt.hardware [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 748.888018] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0f6cd1-f664-4c59-8279-5c18a3f49cf7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.899084] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60fa9842-ab59-4f07-a80f-a8978586abf8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.055824] env[61906]: DEBUG nova.network.neutron [-] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.273176] env[61906]: DEBUG nova.network.neutron [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.389345] env[61906]: DEBUG nova.network.neutron [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.514184] env[61906]: ERROR nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cd611107-eb27-4ac8-b1f6-9a29c1516ac7, please check neutron logs for more information. [ 749.514184] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 749.514184] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.514184] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 749.514184] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.514184] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 749.514184] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.514184] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 749.514184] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.514184] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 749.514184] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.514184] env[61906]: ERROR nova.compute.manager raise self.value [ 749.514184] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.514184] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 749.514184] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.514184] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 749.514602] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.514602] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 749.514602] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cd611107-eb27-4ac8-b1f6-9a29c1516ac7, please check neutron logs for more information. [ 749.514602] env[61906]: ERROR nova.compute.manager [ 749.514602] env[61906]: Traceback (most recent call last): [ 749.514602] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 749.514602] env[61906]: listener.cb(fileno) [ 749.514602] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.514602] env[61906]: result = function(*args, **kwargs) [ 749.514602] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.514602] env[61906]: return func(*args, **kwargs) [ 749.514602] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.514602] env[61906]: raise e [ 749.514602] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.514602] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 749.514602] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.514602] env[61906]: created_port_ids = self._update_ports_for_instance( [ 749.514602] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.514602] env[61906]: with excutils.save_and_reraise_exception(): [ 749.514602] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.514602] env[61906]: self.force_reraise() [ 749.514602] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.514602] env[61906]: raise self.value [ 749.514602] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.514602] env[61906]: updated_port = self._update_port( [ 749.514602] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.514602] env[61906]: _ensure_no_port_binding_failure(port) [ 749.514602] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.514602] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 749.515238] env[61906]: nova.exception.PortBindingFailed: Binding failed for port cd611107-eb27-4ac8-b1f6-9a29c1516ac7, please check neutron logs for more information. [ 749.515238] env[61906]: Removing descriptor: 15 [ 749.515238] env[61906]: ERROR nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cd611107-eb27-4ac8-b1f6-9a29c1516ac7, please check neutron logs for more information. [ 749.515238] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Traceback (most recent call last): [ 749.515238] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 749.515238] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] yield resources [ 749.515238] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.515238] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] self.driver.spawn(context, instance, image_meta, [ 749.515238] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 749.515238] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.515238] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.515238] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] vm_ref = self.build_virtual_machine(instance, [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] for vif in network_info: [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] return self._sync_wrapper(fn, *args, **kwargs) [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] self.wait() [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] self[:] = self._gt.wait() [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] return self._exit_event.wait() [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 749.515496] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] result = hub.switch() [ 749.515847] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 749.515847] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] return self.greenlet.switch() [ 749.515847] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.515847] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] result = function(*args, **kwargs) [ 749.515847] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 749.515847] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] return func(*args, **kwargs) [ 749.515847] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.515847] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] raise e [ 749.515847] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.515847] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] nwinfo = self.network_api.allocate_for_instance( [ 749.515847] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 749.515847] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] created_port_ids = self._update_ports_for_instance( [ 749.515847] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 749.516139] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] with excutils.save_and_reraise_exception(): [ 749.516139] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.516139] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] self.force_reraise() [ 749.516139] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.516139] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] raise self.value [ 749.516139] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 749.516139] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] updated_port = self._update_port( [ 749.516139] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.516139] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] _ensure_no_port_binding_failure(port) [ 749.516139] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.516139] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] raise exception.PortBindingFailed(port_id=port['id']) [ 749.516139] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] nova.exception.PortBindingFailed: Binding failed for port cd611107-eb27-4ac8-b1f6-9a29c1516ac7, please check neutron logs for more information. [ 749.516139] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] [ 749.516448] env[61906]: INFO nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Terminating instance [ 749.517674] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Acquiring lock "refresh_cache-ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.517833] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Acquired lock "refresh_cache-ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.517995] env[61906]: DEBUG nova.network.neutron [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 749.559201] env[61906]: INFO nova.compute.manager [-] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Took 1.03 seconds to deallocate network for instance. [ 749.561484] env[61906]: DEBUG nova.compute.claims [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 749.561678] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.892185] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Releasing lock "refresh_cache-83c86407-16b6-4025-9568-0c9c73d20b10" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.892492] env[61906]: DEBUG nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 749.892753] env[61906]: DEBUG nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 749.892930] env[61906]: DEBUG nova.network.neutron [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 749.916458] env[61906]: DEBUG nova.network.neutron [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.058219] env[61906]: DEBUG nova.network.neutron [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.157439] env[61906]: DEBUG nova.network.neutron [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.222023] env[61906]: DEBUG nova.compute.manager [req-ee7b4aa1-47a4-48d5-9fce-3bf04f46e2b0 req-ebab6724-6557-4b85-9264-d5f33a25b64a service nova] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Received event network-changed-cd611107-eb27-4ac8-b1f6-9a29c1516ac7 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 750.222023] env[61906]: DEBUG nova.compute.manager [req-ee7b4aa1-47a4-48d5-9fce-3bf04f46e2b0 req-ebab6724-6557-4b85-9264-d5f33a25b64a service nova] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Refreshing instance network info cache due to event network-changed-cd611107-eb27-4ac8-b1f6-9a29c1516ac7. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 750.222023] env[61906]: DEBUG oslo_concurrency.lockutils [req-ee7b4aa1-47a4-48d5-9fce-3bf04f46e2b0 req-ebab6724-6557-4b85-9264-d5f33a25b64a service nova] Acquiring lock "refresh_cache-ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.302833] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 83c86407-16b6-4025-9568-0c9c73d20b10 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 750.302833] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 750.302833] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance e0ea5eb3-82cf-44f1-91ee-2c2861aefee4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 750.302833] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance ce3c082c-a3cb-452d-9581-0bed6aea892b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 750.302981] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance a8890726-6c64-4e09-a9a4-0726aa2e8c31 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 750.302981] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 750.416685] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "e133605d-c630-4b9e-a314-bf496c853710" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.417113] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.421991] env[61906]: DEBUG nova.network.neutron [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.658849] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Releasing lock "refresh_cache-ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.660842] env[61906]: DEBUG nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 750.660842] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 750.660842] env[61906]: DEBUG oslo_concurrency.lockutils [req-ee7b4aa1-47a4-48d5-9fce-3bf04f46e2b0 req-ebab6724-6557-4b85-9264-d5f33a25b64a service nova] Acquired lock "refresh_cache-ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.660842] env[61906]: DEBUG nova.network.neutron [req-ee7b4aa1-47a4-48d5-9fce-3bf04f46e2b0 req-ebab6724-6557-4b85-9264-d5f33a25b64a service nova] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Refreshing network info cache for port cd611107-eb27-4ac8-b1f6-9a29c1516ac7 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 750.662798] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5e07d78-28a6-40c2-81fb-f2b091223fa1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.672438] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47758de5-9df9-4c65-a820-4100766b4f75 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.698703] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb could not be found. [ 750.698931] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 750.699152] env[61906]: INFO nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 750.699397] env[61906]: DEBUG oslo.service.loopingcall [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 750.699615] env[61906]: DEBUG nova.compute.manager [-] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 750.699706] env[61906]: DEBUG nova.network.neutron [-] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 750.720654] env[61906]: DEBUG nova.network.neutron [-] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.808413] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 45621a25-cd9c-4931-899c-647da0b50ec7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 750.925717] env[61906]: INFO nova.compute.manager [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] [instance: 83c86407-16b6-4025-9568-0c9c73d20b10] Took 1.03 seconds to deallocate network for instance. [ 751.199037] env[61906]: DEBUG nova.network.neutron [req-ee7b4aa1-47a4-48d5-9fce-3bf04f46e2b0 req-ebab6724-6557-4b85-9264-d5f33a25b64a service nova] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.223443] env[61906]: DEBUG nova.network.neutron [-] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.306906] env[61906]: DEBUG nova.network.neutron [req-ee7b4aa1-47a4-48d5-9fce-3bf04f46e2b0 req-ebab6724-6557-4b85-9264-d5f33a25b64a service nova] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.310065] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance db4eb67a-12dc-4347-ba29-1af46626a87d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 751.672742] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Acquiring lock "0552e012-edcf-444c-a9b3-f1cea93b9e82" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.672979] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Lock "0552e012-edcf-444c-a9b3-f1cea93b9e82" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.725382] env[61906]: INFO nova.compute.manager [-] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Took 1.03 seconds to deallocate network for instance. [ 751.727668] env[61906]: DEBUG nova.compute.claims [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 751.727853] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.809656] env[61906]: DEBUG oslo_concurrency.lockutils [req-ee7b4aa1-47a4-48d5-9fce-3bf04f46e2b0 req-ebab6724-6557-4b85-9264-d5f33a25b64a service nova] Releasing lock "refresh_cache-ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.809656] env[61906]: DEBUG nova.compute.manager [req-ee7b4aa1-47a4-48d5-9fce-3bf04f46e2b0 req-ebab6724-6557-4b85-9264-d5f33a25b64a service nova] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Received event network-vif-deleted-cd611107-eb27-4ac8-b1f6-9a29c1516ac7 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.815224] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance a6e54432-a58e-49ba-a5c8-e6188f2ea4ff has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 752.035452] env[61906]: INFO nova.scheduler.client.report [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Deleted allocations for instance 83c86407-16b6-4025-9568-0c9c73d20b10 [ 752.316804] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance b1429121-3c7e-4982-9059-5eb98a7f92df has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 752.546968] env[61906]: DEBUG oslo_concurrency.lockutils [None req-64857d0c-6514-4512-939e-f6e69fea1b0e tempest-ServersNegativeTestJSON-1396988084 tempest-ServersNegativeTestJSON-1396988084-project-member] Lock "83c86407-16b6-4025-9568-0c9c73d20b10" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.701s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.819786] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance c3cb1a2d-e905-497c-a4de-8c6579c37156 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 753.050065] env[61906]: DEBUG nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 753.323872] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 77791983-cef1-4f07-9b62-d52d335b08c8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 753.573021] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.828273] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 7943ab1a-82df-48dd-874c-15dc015db51e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 754.334019] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 754.834521] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 8ab6b473-d13c-4341-9789-992ac3aba6a2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.338032] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 9a23071d-40fd-4446-aa03-ecbddcafe3d6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.841409] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 5adaa660-b736-4c11-9141-846cf475ccd5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.344772] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.847841] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 79452791-59cb-4722-bb4a-8e59d8c4e641 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.352102] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 7bae658e-b050-4639-b34e-c2671ef5c773 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.352407] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 757.352594] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 757.592427] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2464d2d3-ef5f-48c6-b18d-052f3fe15495 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.600248] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77341ac9-92ec-4e38-a87a-ed79f337562c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.376113] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8dfe27e-4979-4d93-9f3d-1d92c47111c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.381658] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b249f541-406c-4a18-8731-142082cc866e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.396147] env[61906]: DEBUG nova.compute.provider_tree [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.900109] env[61906]: DEBUG nova.scheduler.client.report [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.405149] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 759.405446] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.658s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.405682] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.936s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.408572] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.408733] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Cleaning up deleted instances {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 759.912201] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] There are 3 instances to clean {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 759.912451] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: e23fd777-66f7-4ee1-bdcb-fe19c55f955c] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 760.143041] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dabfce4-7a0e-4505-8836-dfaed6a210e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.150306] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1575fbc-799e-418a-9167-62b390300ee2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.179677] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9729aaa-ee59-457d-a645-6b3c986ba1ed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.187220] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9aa18a-3086-47ff-8849-af0ce02c47fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.200925] env[61906]: DEBUG nova.compute.provider_tree [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.419636] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 7f4bb9fb-fa64-451d-9876-7e4314c13a64] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 760.704496] env[61906]: DEBUG nova.scheduler.client.report [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.921602] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 85effb81-494c-4bda-a093-6dde523d5f45] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 761.209486] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.804s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.210148] env[61906]: ERROR nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c4140c45-ff57-41b0-b86f-24a25f5f62ba, please check neutron logs for more information. [ 761.210148] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Traceback (most recent call last): [ 761.210148] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.210148] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] self.driver.spawn(context, instance, image_meta, [ 761.210148] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 761.210148] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.210148] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.210148] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] vm_ref = self.build_virtual_machine(instance, [ 761.210148] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.210148] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.210148] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.210395] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] for vif in network_info: [ 761.210395] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.210395] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] return self._sync_wrapper(fn, *args, **kwargs) [ 761.210395] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.210395] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] self.wait() [ 761.210395] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.210395] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] self[:] = self._gt.wait() [ 761.210395] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.210395] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] return self._exit_event.wait() [ 761.210395] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 761.210395] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] result = hub.switch() [ 761.210395] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 761.210395] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] return self.greenlet.switch() [ 761.210662] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.210662] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] result = function(*args, **kwargs) [ 761.210662] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 761.210662] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] return func(*args, **kwargs) [ 761.210662] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.210662] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] raise e [ 761.210662] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.210662] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] nwinfo = self.network_api.allocate_for_instance( [ 761.210662] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.210662] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] created_port_ids = self._update_ports_for_instance( [ 761.210662] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.210662] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] with excutils.save_and_reraise_exception(): [ 761.210662] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.210922] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] self.force_reraise() [ 761.210922] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.210922] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] raise self.value [ 761.210922] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.210922] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] updated_port = self._update_port( [ 761.210922] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.210922] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] _ensure_no_port_binding_failure(port) [ 761.210922] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.210922] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] raise exception.PortBindingFailed(port_id=port['id']) [ 761.210922] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] nova.exception.PortBindingFailed: Binding failed for port c4140c45-ff57-41b0-b86f-24a25f5f62ba, please check neutron logs for more information. [ 761.210922] env[61906]: ERROR nova.compute.manager [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] [ 761.211160] env[61906]: DEBUG nova.compute.utils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Binding failed for port c4140c45-ff57-41b0-b86f-24a25f5f62ba, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 761.212065] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.942s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.213603] env[61906]: INFO nova.compute.claims [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.216795] env[61906]: DEBUG nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Build of instance 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9 was re-scheduled: Binding failed for port c4140c45-ff57-41b0-b86f-24a25f5f62ba, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 761.216898] env[61906]: DEBUG nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 761.217070] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Acquiring lock "refresh_cache-2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.217222] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Acquired lock "refresh_cache-2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.217380] env[61906]: DEBUG nova.network.neutron [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.425249] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.425575] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Cleaning up deleted instances with incomplete migration {{(pid=61906) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 761.733616] env[61906]: DEBUG nova.network.neutron [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.823463] env[61906]: DEBUG nova.network.neutron [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.928221] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.325674] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Releasing lock "refresh_cache-2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.325900] env[61906]: DEBUG nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 762.326092] env[61906]: DEBUG nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.326258] env[61906]: DEBUG nova.network.neutron [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 762.342480] env[61906]: DEBUG nova.network.neutron [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.461017] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead17fdb-2539-40e7-bf62-e58e98344741 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.467156] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24df39dd-06e9-4b93-b30a-b63a2c047f59 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.497149] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3664f7b4-2477-47bc-9be1-22e907867c8c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.504851] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2349abf7-17b5-4917-ab46-34c44ce50082 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.518026] env[61906]: DEBUG nova.compute.provider_tree [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.844618] env[61906]: DEBUG nova.network.neutron [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.020630] env[61906]: DEBUG nova.scheduler.client.report [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.347040] env[61906]: INFO nova.compute.manager [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] [instance: 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9] Took 1.02 seconds to deallocate network for instance. [ 763.525754] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.314s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.526161] env[61906]: DEBUG nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 763.528876] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.609s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.034781] env[61906]: DEBUG nova.compute.utils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 764.039027] env[61906]: DEBUG nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 764.039500] env[61906]: DEBUG nova.network.neutron [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 764.075915] env[61906]: DEBUG nova.policy [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '10c96c5d29ca4935bff89789607bf86e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ac2f0981617413b967c9c3ed2b4cd4c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 764.277909] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96b481a-2781-420d-9a7a-df24ba010a9f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.285680] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1976f3-ea49-459f-90c9-3775a55bc8b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.316107] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f22ab2-5e70-4913-98fa-c0ce3cfa9408 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.323547] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d64c140-ad37-4531-9039-3dd66fa5ec78 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.336458] env[61906]: DEBUG nova.compute.provider_tree [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.375342] env[61906]: INFO nova.scheduler.client.report [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Deleted allocations for instance 2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9 [ 764.385887] env[61906]: DEBUG nova.network.neutron [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Successfully created port: b3dbfa76-7b63-45cb-bdbb-6045d381ed93 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 764.542272] env[61906]: DEBUG nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 764.840415] env[61906]: DEBUG nova.scheduler.client.report [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.891033] env[61906]: DEBUG oslo_concurrency.lockutils [None req-01cbd61d-de42-4a78-afe6-c09d0413f864 tempest-FloatingIPsAssociationTestJSON-707516092 tempest-FloatingIPsAssociationTestJSON-707516092-project-member] Lock "2494b3c6-a6a9-4cfe-adf9-28c37d5f71d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.005s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.049901] env[61906]: INFO nova.virt.block_device [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Booting with volume b648a2bf-cab0-4503-9a3e-de117f9b9fc3 at /dev/sda [ 765.088934] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-79dfb416-5029-4325-8b91-70feef5d06b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.097921] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262b5c8d-3b19-4115-9bde-478d16124020 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.118582] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-82807163-8f20-401d-81ba-03c9a79204a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.126013] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c10a2d6-b177-4307-9588-efa49d9cc5bf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.146802] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb247aa-2889-4a08-807f-7be69ae2f064 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.153019] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff24ef0-95c2-44fe-84a0-55ea3da69552 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.166515] env[61906]: DEBUG nova.virt.block_device [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Updating existing volume attachment record: 4eff6199-7f44-4f7c-8c42-0f42d6fba5e0 {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 765.352236] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.823s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.352924] env[61906]: ERROR nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 455aa184-2c4a-450f-b1f4-4288b690df03, please check neutron logs for more information. [ 765.352924] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Traceback (most recent call last): [ 765.352924] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 765.352924] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] self.driver.spawn(context, instance, image_meta, [ 765.352924] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 765.352924] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 765.352924] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 765.352924] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] vm_ref = self.build_virtual_machine(instance, [ 765.352924] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 765.352924] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] vif_infos = vmwarevif.get_vif_info(self._session, [ 765.352924] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] for vif in network_info: [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] return self._sync_wrapper(fn, *args, **kwargs) [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] self.wait() [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] self[:] = self._gt.wait() [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] return self._exit_event.wait() [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] current.throw(*self._exc) [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] result = function(*args, **kwargs) [ 765.353249] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 765.353565] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] return func(*args, **kwargs) [ 765.353565] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.353565] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] raise e [ 765.353565] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.353565] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] nwinfo = self.network_api.allocate_for_instance( [ 765.353565] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.353565] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] created_port_ids = self._update_ports_for_instance( [ 765.353565] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.353565] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] with excutils.save_and_reraise_exception(): [ 765.353565] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.353565] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] self.force_reraise() [ 765.353565] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.353565] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] raise self.value [ 765.353856] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.353856] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] updated_port = self._update_port( [ 765.353856] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.353856] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] _ensure_no_port_binding_failure(port) [ 765.353856] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.353856] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] raise exception.PortBindingFailed(port_id=port['id']) [ 765.353856] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] nova.exception.PortBindingFailed: Binding failed for port 455aa184-2c4a-450f-b1f4-4288b690df03, please check neutron logs for more information. [ 765.353856] env[61906]: ERROR nova.compute.manager [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] [ 765.355358] env[61906]: DEBUG nova.compute.utils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Binding failed for port 455aa184-2c4a-450f-b1f4-4288b690df03, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 765.355746] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.636s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.364019] env[61906]: DEBUG nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Build of instance e0ea5eb3-82cf-44f1-91ee-2c2861aefee4 was re-scheduled: Binding failed for port 455aa184-2c4a-450f-b1f4-4288b690df03, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 765.364019] env[61906]: DEBUG nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 765.364019] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Acquiring lock "refresh_cache-e0ea5eb3-82cf-44f1-91ee-2c2861aefee4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.364019] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Acquired lock "refresh_cache-e0ea5eb3-82cf-44f1-91ee-2c2861aefee4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.364233] env[61906]: DEBUG nova.network.neutron [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 765.366561] env[61906]: DEBUG nova.compute.manager [req-49da2c8e-422b-48ae-acd4-248db4b3eb91 req-ed9617fa-29a6-4a87-a69d-a2d8dd7e9cdb service nova] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Received event network-changed-b3dbfa76-7b63-45cb-bdbb-6045d381ed93 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 765.366858] env[61906]: DEBUG nova.compute.manager [req-49da2c8e-422b-48ae-acd4-248db4b3eb91 req-ed9617fa-29a6-4a87-a69d-a2d8dd7e9cdb service nova] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Refreshing instance network info cache due to event network-changed-b3dbfa76-7b63-45cb-bdbb-6045d381ed93. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 765.367103] env[61906]: DEBUG oslo_concurrency.lockutils [req-49da2c8e-422b-48ae-acd4-248db4b3eb91 req-ed9617fa-29a6-4a87-a69d-a2d8dd7e9cdb service nova] Acquiring lock "refresh_cache-45621a25-cd9c-4931-899c-647da0b50ec7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.367274] env[61906]: DEBUG oslo_concurrency.lockutils [req-49da2c8e-422b-48ae-acd4-248db4b3eb91 req-ed9617fa-29a6-4a87-a69d-a2d8dd7e9cdb service nova] Acquired lock "refresh_cache-45621a25-cd9c-4931-899c-647da0b50ec7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.367459] env[61906]: DEBUG nova.network.neutron [req-49da2c8e-422b-48ae-acd4-248db4b3eb91 req-ed9617fa-29a6-4a87-a69d-a2d8dd7e9cdb service nova] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Refreshing network info cache for port b3dbfa76-7b63-45cb-bdbb-6045d381ed93 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 765.391942] env[61906]: DEBUG nova.compute.manager [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 765.499775] env[61906]: ERROR nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b3dbfa76-7b63-45cb-bdbb-6045d381ed93, please check neutron logs for more information. [ 765.499775] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 765.499775] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.499775] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 765.499775] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.499775] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 765.499775] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.499775] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 765.499775] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.499775] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 765.499775] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.499775] env[61906]: ERROR nova.compute.manager raise self.value [ 765.499775] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.499775] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 765.499775] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.499775] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 765.500147] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.500147] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 765.500147] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b3dbfa76-7b63-45cb-bdbb-6045d381ed93, please check neutron logs for more information. [ 765.500147] env[61906]: ERROR nova.compute.manager [ 765.500147] env[61906]: Traceback (most recent call last): [ 765.500147] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 765.500147] env[61906]: listener.cb(fileno) [ 765.500147] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.500147] env[61906]: result = function(*args, **kwargs) [ 765.500147] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 765.500147] env[61906]: return func(*args, **kwargs) [ 765.500147] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.500147] env[61906]: raise e [ 765.500147] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.500147] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 765.500147] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.500147] env[61906]: created_port_ids = self._update_ports_for_instance( [ 765.500147] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.500147] env[61906]: with excutils.save_and_reraise_exception(): [ 765.500147] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.500147] env[61906]: self.force_reraise() [ 765.500147] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.500147] env[61906]: raise self.value [ 765.500147] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.500147] env[61906]: updated_port = self._update_port( [ 765.500147] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.500147] env[61906]: _ensure_no_port_binding_failure(port) [ 765.500147] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.500147] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 765.500728] env[61906]: nova.exception.PortBindingFailed: Binding failed for port b3dbfa76-7b63-45cb-bdbb-6045d381ed93, please check neutron logs for more information. [ 765.500728] env[61906]: Removing descriptor: 15 [ 765.892878] env[61906]: DEBUG nova.network.neutron [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.899671] env[61906]: DEBUG nova.network.neutron [req-49da2c8e-422b-48ae-acd4-248db4b3eb91 req-ed9617fa-29a6-4a87-a69d-a2d8dd7e9cdb service nova] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.913639] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.983341] env[61906]: DEBUG nova.network.neutron [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.046361] env[61906]: DEBUG nova.network.neutron [req-49da2c8e-422b-48ae-acd4-248db4b3eb91 req-ed9617fa-29a6-4a87-a69d-a2d8dd7e9cdb service nova] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.155041] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8feb6440-dd5c-4049-a9c8-832e9e47ce41 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.161269] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976ab20e-951b-4891-bd09-74f000713bcf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.193212] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbbd9ac7-5b18-486a-8fdc-138e5d200c9f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.203500] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f5a05f3-13f1-4462-80f5-a56005cd7a2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.217661] env[61906]: DEBUG nova.compute.provider_tree [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.486384] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Releasing lock "refresh_cache-e0ea5eb3-82cf-44f1-91ee-2c2861aefee4" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.486662] env[61906]: DEBUG nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 766.486830] env[61906]: DEBUG nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.486995] env[61906]: DEBUG nova.network.neutron [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.506226] env[61906]: DEBUG nova.network.neutron [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.549035] env[61906]: DEBUG oslo_concurrency.lockutils [req-49da2c8e-422b-48ae-acd4-248db4b3eb91 req-ed9617fa-29a6-4a87-a69d-a2d8dd7e9cdb service nova] Releasing lock "refresh_cache-45621a25-cd9c-4931-899c-647da0b50ec7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.720886] env[61906]: DEBUG nova.scheduler.client.report [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.009752] env[61906]: DEBUG nova.network.neutron [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.225893] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.870s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.226631] env[61906]: ERROR nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port be810f4e-b293-4b85-b0dc-133b33010e8a, please check neutron logs for more information. [ 767.226631] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Traceback (most recent call last): [ 767.226631] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 767.226631] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] self.driver.spawn(context, instance, image_meta, [ 767.226631] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 767.226631] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.226631] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.226631] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] vm_ref = self.build_virtual_machine(instance, [ 767.226631] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.226631] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.226631] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.226900] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] for vif in network_info: [ 767.226900] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 767.226900] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] return self._sync_wrapper(fn, *args, **kwargs) [ 767.226900] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 767.226900] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] self.wait() [ 767.226900] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 767.226900] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] self[:] = self._gt.wait() [ 767.226900] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.226900] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] return self._exit_event.wait() [ 767.226900] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 767.226900] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] result = hub.switch() [ 767.226900] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 767.226900] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] return self.greenlet.switch() [ 767.227416] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.227416] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] result = function(*args, **kwargs) [ 767.227416] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 767.227416] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] return func(*args, **kwargs) [ 767.227416] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.227416] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] raise e [ 767.227416] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.227416] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] nwinfo = self.network_api.allocate_for_instance( [ 767.227416] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.227416] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] created_port_ids = self._update_ports_for_instance( [ 767.227416] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.227416] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] with excutils.save_and_reraise_exception(): [ 767.227416] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.227842] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] self.force_reraise() [ 767.227842] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.227842] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] raise self.value [ 767.227842] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.227842] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] updated_port = self._update_port( [ 767.227842] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.227842] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] _ensure_no_port_binding_failure(port) [ 767.227842] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.227842] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] raise exception.PortBindingFailed(port_id=port['id']) [ 767.227842] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] nova.exception.PortBindingFailed: Binding failed for port be810f4e-b293-4b85-b0dc-133b33010e8a, please check neutron logs for more information. [ 767.227842] env[61906]: ERROR nova.compute.manager [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] [ 767.228189] env[61906]: DEBUG nova.compute.utils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Binding failed for port be810f4e-b293-4b85-b0dc-133b33010e8a, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 767.229575] env[61906]: DEBUG nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Build of instance ce3c082c-a3cb-452d-9581-0bed6aea892b was re-scheduled: Binding failed for port be810f4e-b293-4b85-b0dc-133b33010e8a, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 767.230719] env[61906]: DEBUG nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 767.230719] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "refresh_cache-ce3c082c-a3cb-452d-9581-0bed6aea892b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.230719] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "refresh_cache-ce3c082c-a3cb-452d-9581-0bed6aea892b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.230904] env[61906]: DEBUG nova.network.neutron [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.231935] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.256s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.233657] env[61906]: INFO nova.compute.claims [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 767.266051] env[61906]: DEBUG nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 767.266668] env[61906]: DEBUG nova.virt.hardware [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 767.266830] env[61906]: DEBUG nova.virt.hardware [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 767.266980] env[61906]: DEBUG nova.virt.hardware [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 767.267206] env[61906]: DEBUG nova.virt.hardware [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 767.267807] env[61906]: DEBUG nova.virt.hardware [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 767.267807] env[61906]: DEBUG nova.virt.hardware [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 767.267807] env[61906]: DEBUG nova.virt.hardware [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 767.269964] env[61906]: DEBUG nova.virt.hardware [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 767.269964] env[61906]: DEBUG nova.virt.hardware [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 767.269964] env[61906]: DEBUG nova.virt.hardware [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 767.269964] env[61906]: DEBUG nova.virt.hardware [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.271329] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08eb86c9-4fe4-4552-ba79-9ad6f257d0a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.280824] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4cb387-c806-4ba8-8ca2-f342a9701702 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.296220] env[61906]: ERROR nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b3dbfa76-7b63-45cb-bdbb-6045d381ed93, please check neutron logs for more information. [ 767.296220] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Traceback (most recent call last): [ 767.296220] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 767.296220] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] yield resources [ 767.296220] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 767.296220] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] self.driver.spawn(context, instance, image_meta, [ 767.296220] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 767.296220] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.296220] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.296220] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] vm_ref = self.build_virtual_machine(instance, [ 767.296220] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.296510] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.296510] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.296510] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] for vif in network_info: [ 767.296510] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 767.296510] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] return self._sync_wrapper(fn, *args, **kwargs) [ 767.296510] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 767.296510] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] self.wait() [ 767.296510] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 767.296510] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] self[:] = self._gt.wait() [ 767.296510] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.296510] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] return self._exit_event.wait() [ 767.296510] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 767.296510] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] current.throw(*self._exc) [ 767.296840] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.296840] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] result = function(*args, **kwargs) [ 767.296840] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 767.296840] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] return func(*args, **kwargs) [ 767.296840] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.296840] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] raise e [ 767.296840] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.296840] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] nwinfo = self.network_api.allocate_for_instance( [ 767.296840] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.296840] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] created_port_ids = self._update_ports_for_instance( [ 767.296840] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.296840] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] with excutils.save_and_reraise_exception(): [ 767.296840] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.297142] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] self.force_reraise() [ 767.297142] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.297142] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] raise self.value [ 767.297142] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.297142] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] updated_port = self._update_port( [ 767.297142] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.297142] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] _ensure_no_port_binding_failure(port) [ 767.297142] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.297142] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] raise exception.PortBindingFailed(port_id=port['id']) [ 767.297142] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] nova.exception.PortBindingFailed: Binding failed for port b3dbfa76-7b63-45cb-bdbb-6045d381ed93, please check neutron logs for more information. [ 767.297142] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] [ 767.297142] env[61906]: INFO nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Terminating instance [ 767.298493] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Acquiring lock "refresh_cache-45621a25-cd9c-4931-899c-647da0b50ec7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.298927] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Acquired lock "refresh_cache-45621a25-cd9c-4931-899c-647da0b50ec7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.298927] env[61906]: DEBUG nova.network.neutron [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.385469] env[61906]: DEBUG nova.compute.manager [req-e85c42cd-ffd0-4d13-b507-496b8086f854 req-74bc1f4a-652a-4098-a908-ef9c38d541a6 service nova] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Received event network-vif-deleted-b3dbfa76-7b63-45cb-bdbb-6045d381ed93 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 767.514865] env[61906]: INFO nova.compute.manager [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] [instance: e0ea5eb3-82cf-44f1-91ee-2c2861aefee4] Took 1.03 seconds to deallocate network for instance. [ 767.759900] env[61906]: DEBUG nova.network.neutron [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.816617] env[61906]: DEBUG nova.network.neutron [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.853037] env[61906]: DEBUG nova.network.neutron [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.934409] env[61906]: DEBUG nova.network.neutron [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.355971] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "refresh_cache-ce3c082c-a3cb-452d-9581-0bed6aea892b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.356234] env[61906]: DEBUG nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 768.356414] env[61906]: DEBUG nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.357419] env[61906]: DEBUG nova.network.neutron [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 768.372371] env[61906]: DEBUG nova.network.neutron [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.437125] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Releasing lock "refresh_cache-45621a25-cd9c-4931-899c-647da0b50ec7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.437668] env[61906]: DEBUG nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 768.437985] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f464f1c6-e7a1-44ff-a91f-b4e9de31a1ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.449292] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3412ef40-95cd-471e-9d94-6bb00b9e330a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.473240] env[61906]: WARNING nova.virt.vmwareapi.driver [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 45621a25-cd9c-4931-899c-647da0b50ec7 could not be found. [ 768.473423] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.475885] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e3697207-65d3-494a-b6a7-5c5fa695b82e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.483668] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f41587ec-d0c5-40e1-aa00-fa05eaa30d88 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.507696] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 45621a25-cd9c-4931-899c-647da0b50ec7 could not be found. [ 768.507934] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 768.508129] env[61906]: INFO nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Took 0.07 seconds to destroy the instance on the hypervisor. [ 768.508375] env[61906]: DEBUG oslo.service.loopingcall [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.509335] env[61906]: DEBUG nova.compute.manager [-] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.509434] env[61906]: DEBUG nova.network.neutron [-] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 768.511628] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ff9ddd-aa58-423a-9b87-ff55de4b3662 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.519110] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d387e62-82e8-4ed9-9bb6-f95d02805a3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.552921] env[61906]: DEBUG nova.network.neutron [-] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.555799] env[61906]: INFO nova.scheduler.client.report [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Deleted allocations for instance e0ea5eb3-82cf-44f1-91ee-2c2861aefee4 [ 768.561566] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f84ae181-158b-48f5-98eb-8ff861b7c01d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.571083] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cdd33c1-6dfe-42b5-aa73-dfcdb936844d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.585405] env[61906]: DEBUG nova.compute.provider_tree [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.875073] env[61906]: DEBUG nova.network.neutron [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.055822] env[61906]: DEBUG nova.network.neutron [-] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.066260] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6d53138-fd99-4e16-adfd-35bfb3f8ed42 tempest-ServersV294TestFqdnHostnames-215477774 tempest-ServersV294TestFqdnHostnames-215477774-project-member] Lock "e0ea5eb3-82cf-44f1-91ee-2c2861aefee4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.186s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.089264] env[61906]: DEBUG nova.scheduler.client.report [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.378070] env[61906]: INFO nova.compute.manager [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: ce3c082c-a3cb-452d-9581-0bed6aea892b] Took 1.02 seconds to deallocate network for instance. [ 769.558827] env[61906]: INFO nova.compute.manager [-] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Took 1.05 seconds to deallocate network for instance. [ 769.568677] env[61906]: DEBUG nova.compute.manager [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.595288] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.362s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.595288] env[61906]: DEBUG nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 769.600145] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.755s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.600145] env[61906]: INFO nova.compute.claims [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 770.099548] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.107955] env[61906]: DEBUG nova.compute.utils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 770.110439] env[61906]: DEBUG nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 770.110730] env[61906]: DEBUG nova.network.neutron [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 770.126679] env[61906]: INFO nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Took 0.57 seconds to detach 1 volumes for instance. [ 770.131033] env[61906]: DEBUG nova.compute.claims [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 770.131143] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.156160] env[61906]: DEBUG nova.policy [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '493de226fb2942f4be9d73472540013b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4f36e1be9fa4798be55371e39e15382', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 770.438211] env[61906]: INFO nova.scheduler.client.report [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleted allocations for instance ce3c082c-a3cb-452d-9581-0bed6aea892b [ 770.605947] env[61906]: DEBUG nova.network.neutron [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Successfully created port: 12ff7474-0485-4fe4-9e38-6cc050ac31f8 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 770.610822] env[61906]: DEBUG nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 770.913738] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9b2bca-fc8c-43a3-9aad-efedf8eadc86 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.921808] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36336491-200c-4dbf-9bd5-e12b026df682 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.955093] env[61906]: DEBUG oslo_concurrency.lockutils [None req-18664c56-ae86-4898-b8e9-e49fe5e7b45b tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "ce3c082c-a3cb-452d-9581-0bed6aea892b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.272s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.957089] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b54622b-5d45-4c3a-a7ca-089bc143f90d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.965379] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a05e494-5a7e-4b37-b66f-8d513d9acb07 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.981034] env[61906]: DEBUG nova.compute.provider_tree [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.460931] env[61906]: DEBUG nova.compute.manager [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 771.484068] env[61906]: DEBUG nova.scheduler.client.report [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.629274] env[61906]: DEBUG nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 771.659948] env[61906]: DEBUG nova.virt.hardware [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 771.659948] env[61906]: DEBUG nova.virt.hardware [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 771.659948] env[61906]: DEBUG nova.virt.hardware [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 771.663697] env[61906]: DEBUG nova.virt.hardware [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 771.663697] env[61906]: DEBUG nova.virt.hardware [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 771.663697] env[61906]: DEBUG nova.virt.hardware [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 771.663697] env[61906]: DEBUG nova.virt.hardware [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 771.663697] env[61906]: DEBUG nova.virt.hardware [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 771.664053] env[61906]: DEBUG nova.virt.hardware [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 771.664053] env[61906]: DEBUG nova.virt.hardware [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 771.664053] env[61906]: DEBUG nova.virt.hardware [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.664053] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3df701a-9a72-4d10-8c9f-5463758fce38 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.673899] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f6b872-7e0c-4128-9ecc-b384a4a646a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.983617] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.991040] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.991040] env[61906]: DEBUG nova.compute.manager [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 771.994459] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.187s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.996355] env[61906]: INFO nova.compute.claims [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.007382] env[61906]: DEBUG nova.compute.manager [req-ebf44330-9859-4101-bafb-7da7377c3b10 req-a82db52d-6f4a-46ff-a0a2-854be472b43e service nova] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Received event network-changed-12ff7474-0485-4fe4-9e38-6cc050ac31f8 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.007675] env[61906]: DEBUG nova.compute.manager [req-ebf44330-9859-4101-bafb-7da7377c3b10 req-a82db52d-6f4a-46ff-a0a2-854be472b43e service nova] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Refreshing instance network info cache due to event network-changed-12ff7474-0485-4fe4-9e38-6cc050ac31f8. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 772.008906] env[61906]: DEBUG oslo_concurrency.lockutils [req-ebf44330-9859-4101-bafb-7da7377c3b10 req-a82db52d-6f4a-46ff-a0a2-854be472b43e service nova] Acquiring lock "refresh_cache-db4eb67a-12dc-4347-ba29-1af46626a87d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.008906] env[61906]: DEBUG oslo_concurrency.lockutils [req-ebf44330-9859-4101-bafb-7da7377c3b10 req-a82db52d-6f4a-46ff-a0a2-854be472b43e service nova] Acquired lock "refresh_cache-db4eb67a-12dc-4347-ba29-1af46626a87d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.008906] env[61906]: DEBUG nova.network.neutron [req-ebf44330-9859-4101-bafb-7da7377c3b10 req-a82db52d-6f4a-46ff-a0a2-854be472b43e service nova] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Refreshing network info cache for port 12ff7474-0485-4fe4-9e38-6cc050ac31f8 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 772.307658] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "24b4a747-67aa-4388-aed6-cb646cd55765" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.309459] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "24b4a747-67aa-4388-aed6-cb646cd55765" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.363932] env[61906]: ERROR nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 12ff7474-0485-4fe4-9e38-6cc050ac31f8, please check neutron logs for more information. [ 772.363932] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 772.363932] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.363932] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 772.363932] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.363932] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 772.363932] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.363932] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 772.363932] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.363932] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 772.363932] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.363932] env[61906]: ERROR nova.compute.manager raise self.value [ 772.363932] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.363932] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 772.363932] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.363932] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 772.364437] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.364437] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 772.364437] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 12ff7474-0485-4fe4-9e38-6cc050ac31f8, please check neutron logs for more information. [ 772.364437] env[61906]: ERROR nova.compute.manager [ 772.364437] env[61906]: Traceback (most recent call last): [ 772.364437] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 772.364437] env[61906]: listener.cb(fileno) [ 772.364437] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.364437] env[61906]: result = function(*args, **kwargs) [ 772.364437] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 772.364437] env[61906]: return func(*args, **kwargs) [ 772.364437] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.364437] env[61906]: raise e [ 772.364437] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.364437] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 772.364437] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.364437] env[61906]: created_port_ids = self._update_ports_for_instance( [ 772.364437] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.364437] env[61906]: with excutils.save_and_reraise_exception(): [ 772.364437] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.364437] env[61906]: self.force_reraise() [ 772.364437] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.364437] env[61906]: raise self.value [ 772.364437] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.364437] env[61906]: updated_port = self._update_port( [ 772.364437] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.364437] env[61906]: _ensure_no_port_binding_failure(port) [ 772.364437] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.364437] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 772.365203] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 12ff7474-0485-4fe4-9e38-6cc050ac31f8, please check neutron logs for more information. [ 772.365203] env[61906]: Removing descriptor: 17 [ 772.365203] env[61906]: ERROR nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 12ff7474-0485-4fe4-9e38-6cc050ac31f8, please check neutron logs for more information. [ 772.365203] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Traceback (most recent call last): [ 772.365203] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 772.365203] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] yield resources [ 772.365203] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 772.365203] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] self.driver.spawn(context, instance, image_meta, [ 772.365203] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 772.365203] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 772.365203] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 772.365203] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] vm_ref = self.build_virtual_machine(instance, [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] vif_infos = vmwarevif.get_vif_info(self._session, [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] for vif in network_info: [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] return self._sync_wrapper(fn, *args, **kwargs) [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] self.wait() [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] self[:] = self._gt.wait() [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] return self._exit_event.wait() [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 772.365531] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] result = hub.switch() [ 772.365892] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 772.365892] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] return self.greenlet.switch() [ 772.365892] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.365892] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] result = function(*args, **kwargs) [ 772.365892] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 772.365892] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] return func(*args, **kwargs) [ 772.365892] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.365892] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] raise e [ 772.365892] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.365892] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] nwinfo = self.network_api.allocate_for_instance( [ 772.365892] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.365892] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] created_port_ids = self._update_ports_for_instance( [ 772.365892] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.366239] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] with excutils.save_and_reraise_exception(): [ 772.366239] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.366239] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] self.force_reraise() [ 772.366239] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.366239] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] raise self.value [ 772.366239] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.366239] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] updated_port = self._update_port( [ 772.366239] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.366239] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] _ensure_no_port_binding_failure(port) [ 772.366239] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.366239] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] raise exception.PortBindingFailed(port_id=port['id']) [ 772.366239] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] nova.exception.PortBindingFailed: Binding failed for port 12ff7474-0485-4fe4-9e38-6cc050ac31f8, please check neutron logs for more information. [ 772.366239] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] [ 772.366626] env[61906]: INFO nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Terminating instance [ 772.367799] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Acquiring lock "refresh_cache-db4eb67a-12dc-4347-ba29-1af46626a87d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.497373] env[61906]: DEBUG nova.compute.utils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.498879] env[61906]: DEBUG nova.compute.manager [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 772.539842] env[61906]: DEBUG nova.network.neutron [req-ebf44330-9859-4101-bafb-7da7377c3b10 req-a82db52d-6f4a-46ff-a0a2-854be472b43e service nova] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.658516] env[61906]: DEBUG nova.network.neutron [req-ebf44330-9859-4101-bafb-7da7377c3b10 req-a82db52d-6f4a-46ff-a0a2-854be472b43e service nova] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.000573] env[61906]: DEBUG nova.compute.manager [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 773.161121] env[61906]: DEBUG oslo_concurrency.lockutils [req-ebf44330-9859-4101-bafb-7da7377c3b10 req-a82db52d-6f4a-46ff-a0a2-854be472b43e service nova] Releasing lock "refresh_cache-db4eb67a-12dc-4347-ba29-1af46626a87d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.163522] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Acquired lock "refresh_cache-db4eb67a-12dc-4347-ba29-1af46626a87d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.163711] env[61906]: DEBUG nova.network.neutron [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.271419] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-706c7a9f-c19d-4eed-880e-aeb39be3913a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.279705] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a8cde5-5339-425a-8c50-ae824f36d0f9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.309019] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05756d0d-ac78-4ae3-880a-2a725c8c42e9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.316548] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2eed2a-26ea-47a1-89ff-51a7ca0ce5ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.330160] env[61906]: DEBUG nova.compute.provider_tree [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.695174] env[61906]: DEBUG nova.network.neutron [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.794206] env[61906]: DEBUG nova.network.neutron [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.836075] env[61906]: DEBUG nova.scheduler.client.report [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.016476] env[61906]: DEBUG nova.compute.manager [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 774.046100] env[61906]: DEBUG nova.compute.manager [req-8584cbe9-5ece-464b-909e-2e40b9e3f332 req-eb53b7bf-adeb-44f4-8719-2ddcd81644c9 service nova] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Received event network-vif-deleted-12ff7474-0485-4fe4-9e38-6cc050ac31f8 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 774.047975] env[61906]: DEBUG nova.virt.hardware [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 774.048366] env[61906]: DEBUG nova.virt.hardware [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 774.048636] env[61906]: DEBUG nova.virt.hardware [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 774.048920] env[61906]: DEBUG nova.virt.hardware [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 774.049179] env[61906]: DEBUG nova.virt.hardware [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 774.049423] env[61906]: DEBUG nova.virt.hardware [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 774.049739] env[61906]: DEBUG nova.virt.hardware [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 774.049996] env[61906]: DEBUG nova.virt.hardware [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 774.050275] env[61906]: DEBUG nova.virt.hardware [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 774.050537] env[61906]: DEBUG nova.virt.hardware [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 774.050808] env[61906]: DEBUG nova.virt.hardware [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 774.052156] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8708af0f-1e2f-4f24-be01-051c4ec5091b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.060826] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7e0eab-15f0-4e22-864f-51154f3b2fb1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.081775] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 774.089391] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Creating folder: Project (4a44b03fd7934d21bd182f5b46115473). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 774.089391] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c26f291b-b4ad-49ad-8742-f5d38c7e14db {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.099726] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Created folder: Project (4a44b03fd7934d21bd182f5b46115473) in parent group-v288914. [ 774.099902] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Creating folder: Instances. Parent ref: group-v288934. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 774.100134] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f874f79-ed63-4af2-971d-56e7dfde8f7e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.109113] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Created folder: Instances in parent group-v288934. [ 774.109345] env[61906]: DEBUG oslo.service.loopingcall [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.109535] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 774.109722] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8676d3d1-3569-4a22-b119-c72784e5c6ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.126065] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 774.126065] env[61906]: value = "task-1356500" [ 774.126065] env[61906]: _type = "Task" [ 774.126065] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.133678] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356500, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.297073] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Releasing lock "refresh_cache-db4eb67a-12dc-4347-ba29-1af46626a87d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.297563] env[61906]: DEBUG nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 774.297864] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 774.298221] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e60e2f67-911f-40cd-8a7b-03396a33cda3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.312903] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1454da-2c20-48ef-a3e4-64cdc63222fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.339623] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.340275] env[61906]: DEBUG nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 774.347259] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.785s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.350502] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance db4eb67a-12dc-4347-ba29-1af46626a87d could not be found. [ 774.350614] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 774.350793] env[61906]: INFO nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 774.351049] env[61906]: DEBUG oslo.service.loopingcall [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.351647] env[61906]: DEBUG nova.compute.manager [-] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.351739] env[61906]: DEBUG nova.network.neutron [-] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 774.374743] env[61906]: DEBUG nova.network.neutron [-] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.637163] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356500, 'name': CreateVM_Task, 'duration_secs': 0.263305} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.640566] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 774.640566] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.640566] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.640566] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 774.640566] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67c25854-d5b1-40c6-ae94-cd0a05f92d5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.644554] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 774.644554] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]524f1fc5-b056-ec66-f631-be51c92109ac" [ 774.644554] env[61906]: _type = "Task" [ 774.644554] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.664444] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]524f1fc5-b056-ec66-f631-be51c92109ac, 'name': SearchDatastore_Task, 'duration_secs': 0.011958} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.664750] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.664975] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 774.665273] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.665452] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.665634] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 774.665882] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3c6d1f8-b745-43ec-ba80-03003656aeb6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.673475] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 774.673651] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 774.674331] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2151a78-087f-4ae7-9104-28685c9ece78 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.681044] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 774.681044] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52bc7e5b-7d3b-b6af-7f81-6ca5be883375" [ 774.681044] env[61906]: _type = "Task" [ 774.681044] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.687384] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52bc7e5b-7d3b-b6af-7f81-6ca5be883375, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.853307] env[61906]: DEBUG nova.compute.utils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.858257] env[61906]: DEBUG nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.861797] env[61906]: DEBUG nova.network.neutron [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 774.881122] env[61906]: DEBUG nova.network.neutron [-] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.923440] env[61906]: DEBUG nova.policy [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca80129d492e4d68b96ef14a63336ef3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb85e88fd4f54e0db2ff131f81137f64', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 775.196572] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52bc7e5b-7d3b-b6af-7f81-6ca5be883375, 'name': SearchDatastore_Task, 'duration_secs': 0.008926} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.197550] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9b13591-4600-4868-8953-00774fdc9821 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.206340] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 775.206340] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]526f0eb4-d188-e55a-fab1-0dcd2a15ca5c" [ 775.206340] env[61906]: _type = "Task" [ 775.206340] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.216679] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]526f0eb4-d188-e55a-fab1-0dcd2a15ca5c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.219781] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab075fb-043d-4c3c-bdad-52070d77fa3a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.227602] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c74774-d8c1-4b5d-92ee-e551bb0d5958 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.265428] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04d93b9-2fa2-46d2-8d12-79433780b945 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.273982] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff275394-2414-474e-a480-caed8919c446 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.289261] env[61906]: DEBUG nova.compute.provider_tree [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.339665] env[61906]: DEBUG nova.network.neutron [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Successfully created port: dd3c4c33-8c85-4958-9053-440584018c37 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 775.358446] env[61906]: DEBUG nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 775.382945] env[61906]: INFO nova.compute.manager [-] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Took 1.03 seconds to deallocate network for instance. [ 775.388138] env[61906]: DEBUG nova.compute.claims [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 775.388422] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.717395] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]526f0eb4-d188-e55a-fab1-0dcd2a15ca5c, 'name': SearchDatastore_Task, 'duration_secs': 0.010065} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.717680] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.717939] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] a6e54432-a58e-49ba-a5c8-e6188f2ea4ff/a6e54432-a58e-49ba-a5c8-e6188f2ea4ff.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 775.718216] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a9974ef-efbd-4064-a73d-16e4c280ad30 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.724961] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 775.724961] env[61906]: value = "task-1356501" [ 775.724961] env[61906]: _type = "Task" [ 775.724961] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.733158] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356501, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.793496] env[61906]: DEBUG nova.scheduler.client.report [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.236873] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356501, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498628} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.237590] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] a6e54432-a58e-49ba-a5c8-e6188f2ea4ff/a6e54432-a58e-49ba-a5c8-e6188f2ea4ff.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 776.237981] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 776.238413] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1892e3d-a78d-4850-bb3a-cf6752c8d311 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.245025] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 776.245025] env[61906]: value = "task-1356502" [ 776.245025] env[61906]: _type = "Task" [ 776.245025] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.256574] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356502, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.299308] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.952s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.300045] env[61906]: ERROR nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 04d25abb-d1a7-496f-9ef8-6dfc268ee592, please check neutron logs for more information. [ 776.300045] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Traceback (most recent call last): [ 776.300045] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 776.300045] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] self.driver.spawn(context, instance, image_meta, [ 776.300045] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 776.300045] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] self._vmops.spawn(context, instance, image_meta, injected_files, [ 776.300045] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 776.300045] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] vm_ref = self.build_virtual_machine(instance, [ 776.300045] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 776.300045] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] vif_infos = vmwarevif.get_vif_info(self._session, [ 776.300045] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] for vif in network_info: [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] return self._sync_wrapper(fn, *args, **kwargs) [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] self.wait() [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] self[:] = self._gt.wait() [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] return self._exit_event.wait() [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] current.throw(*self._exc) [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] result = function(*args, **kwargs) [ 776.300409] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 776.300816] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] return func(*args, **kwargs) [ 776.300816] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.300816] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] raise e [ 776.300816] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.300816] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] nwinfo = self.network_api.allocate_for_instance( [ 776.300816] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.300816] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] created_port_ids = self._update_ports_for_instance( [ 776.300816] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.300816] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] with excutils.save_and_reraise_exception(): [ 776.300816] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.300816] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] self.force_reraise() [ 776.300816] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.300816] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] raise self.value [ 776.301141] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.301141] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] updated_port = self._update_port( [ 776.301141] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.301141] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] _ensure_no_port_binding_failure(port) [ 776.301141] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.301141] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] raise exception.PortBindingFailed(port_id=port['id']) [ 776.301141] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] nova.exception.PortBindingFailed: Binding failed for port 04d25abb-d1a7-496f-9ef8-6dfc268ee592, please check neutron logs for more information. [ 776.301141] env[61906]: ERROR nova.compute.manager [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] [ 776.301141] env[61906]: DEBUG nova.compute.utils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Binding failed for port 04d25abb-d1a7-496f-9ef8-6dfc268ee592, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 776.302140] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.574s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.305217] env[61906]: DEBUG nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Build of instance a8890726-6c64-4e09-a9a4-0726aa2e8c31 was re-scheduled: Binding failed for port 04d25abb-d1a7-496f-9ef8-6dfc268ee592, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 776.305673] env[61906]: DEBUG nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 776.307808] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Acquiring lock "refresh_cache-a8890726-6c64-4e09-a9a4-0726aa2e8c31" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.307808] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Acquired lock "refresh_cache-a8890726-6c64-4e09-a9a4-0726aa2e8c31" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.307808] env[61906]: DEBUG nova.network.neutron [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 776.346319] env[61906]: DEBUG nova.compute.manager [req-d93e9082-68bd-4515-932c-21e7bcc85c81 req-f29b34eb-1479-47e6-8234-2f8a57316c71 service nova] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Received event network-changed-dd3c4c33-8c85-4958-9053-440584018c37 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.346319] env[61906]: DEBUG nova.compute.manager [req-d93e9082-68bd-4515-932c-21e7bcc85c81 req-f29b34eb-1479-47e6-8234-2f8a57316c71 service nova] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Refreshing instance network info cache due to event network-changed-dd3c4c33-8c85-4958-9053-440584018c37. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 776.346319] env[61906]: DEBUG oslo_concurrency.lockutils [req-d93e9082-68bd-4515-932c-21e7bcc85c81 req-f29b34eb-1479-47e6-8234-2f8a57316c71 service nova] Acquiring lock "refresh_cache-b1429121-3c7e-4982-9059-5eb98a7f92df" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.346319] env[61906]: DEBUG oslo_concurrency.lockutils [req-d93e9082-68bd-4515-932c-21e7bcc85c81 req-f29b34eb-1479-47e6-8234-2f8a57316c71 service nova] Acquired lock "refresh_cache-b1429121-3c7e-4982-9059-5eb98a7f92df" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.346319] env[61906]: DEBUG nova.network.neutron [req-d93e9082-68bd-4515-932c-21e7bcc85c81 req-f29b34eb-1479-47e6-8234-2f8a57316c71 service nova] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Refreshing network info cache for port dd3c4c33-8c85-4958-9053-440584018c37 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 776.368488] env[61906]: DEBUG nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 776.398503] env[61906]: DEBUG nova.virt.hardware [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.398760] env[61906]: DEBUG nova.virt.hardware [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.398934] env[61906]: DEBUG nova.virt.hardware [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.399114] env[61906]: DEBUG nova.virt.hardware [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.399298] env[61906]: DEBUG nova.virt.hardware [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.399411] env[61906]: DEBUG nova.virt.hardware [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.400076] env[61906]: DEBUG nova.virt.hardware [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.400135] env[61906]: DEBUG nova.virt.hardware [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.400360] env[61906]: DEBUG nova.virt.hardware [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.400584] env[61906]: DEBUG nova.virt.hardware [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.400842] env[61906]: DEBUG nova.virt.hardware [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.402193] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0ac2c8-5d93-4d0b-b161-ca3a45ae9fa9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.413195] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaeab901-5bc4-4352-89f3-a76ee45431ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.574791] env[61906]: ERROR nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dd3c4c33-8c85-4958-9053-440584018c37, please check neutron logs for more information. [ 776.574791] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 776.574791] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.574791] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 776.574791] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.574791] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 776.574791] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.574791] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 776.574791] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.574791] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 776.574791] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.574791] env[61906]: ERROR nova.compute.manager raise self.value [ 776.574791] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.574791] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 776.574791] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.574791] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 776.575438] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.575438] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 776.575438] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dd3c4c33-8c85-4958-9053-440584018c37, please check neutron logs for more information. [ 776.575438] env[61906]: ERROR nova.compute.manager [ 776.575438] env[61906]: Traceback (most recent call last): [ 776.575438] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 776.575438] env[61906]: listener.cb(fileno) [ 776.575438] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.575438] env[61906]: result = function(*args, **kwargs) [ 776.575438] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 776.575438] env[61906]: return func(*args, **kwargs) [ 776.575438] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.575438] env[61906]: raise e [ 776.575438] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.575438] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 776.575438] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.575438] env[61906]: created_port_ids = self._update_ports_for_instance( [ 776.575438] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.575438] env[61906]: with excutils.save_and_reraise_exception(): [ 776.575438] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.575438] env[61906]: self.force_reraise() [ 776.575438] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.575438] env[61906]: raise self.value [ 776.575438] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.575438] env[61906]: updated_port = self._update_port( [ 776.575438] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.575438] env[61906]: _ensure_no_port_binding_failure(port) [ 776.575438] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.575438] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 776.576173] env[61906]: nova.exception.PortBindingFailed: Binding failed for port dd3c4c33-8c85-4958-9053-440584018c37, please check neutron logs for more information. [ 776.576173] env[61906]: Removing descriptor: 17 [ 776.576173] env[61906]: ERROR nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dd3c4c33-8c85-4958-9053-440584018c37, please check neutron logs for more information. [ 776.576173] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Traceback (most recent call last): [ 776.576173] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 776.576173] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] yield resources [ 776.576173] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 776.576173] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] self.driver.spawn(context, instance, image_meta, [ 776.576173] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 776.576173] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] self._vmops.spawn(context, instance, image_meta, injected_files, [ 776.576173] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 776.576173] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] vm_ref = self.build_virtual_machine(instance, [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] vif_infos = vmwarevif.get_vif_info(self._session, [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] for vif in network_info: [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] return self._sync_wrapper(fn, *args, **kwargs) [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] self.wait() [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] self[:] = self._gt.wait() [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] return self._exit_event.wait() [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 776.576510] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] result = hub.switch() [ 776.577035] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 776.577035] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] return self.greenlet.switch() [ 776.577035] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.577035] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] result = function(*args, **kwargs) [ 776.577035] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 776.577035] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] return func(*args, **kwargs) [ 776.577035] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.577035] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] raise e [ 776.577035] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.577035] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] nwinfo = self.network_api.allocate_for_instance( [ 776.577035] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.577035] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] created_port_ids = self._update_ports_for_instance( [ 776.577035] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.577348] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] with excutils.save_and_reraise_exception(): [ 776.577348] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.577348] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] self.force_reraise() [ 776.577348] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.577348] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] raise self.value [ 776.577348] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.577348] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] updated_port = self._update_port( [ 776.577348] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.577348] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] _ensure_no_port_binding_failure(port) [ 776.577348] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.577348] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] raise exception.PortBindingFailed(port_id=port['id']) [ 776.577348] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] nova.exception.PortBindingFailed: Binding failed for port dd3c4c33-8c85-4958-9053-440584018c37, please check neutron logs for more information. [ 776.577348] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] [ 776.577674] env[61906]: INFO nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Terminating instance [ 776.577674] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "refresh_cache-b1429121-3c7e-4982-9059-5eb98a7f92df" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.757647] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356502, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065714} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.757934] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 776.758755] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9454bf16-6d1a-4f0d-9ab3-baba36a0abe3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.778479] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] a6e54432-a58e-49ba-a5c8-e6188f2ea4ff/a6e54432-a58e-49ba-a5c8-e6188f2ea4ff.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 776.778782] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cdcb1b97-0b0f-4aec-9c1b-f03d49f5122e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.801315] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 776.801315] env[61906]: value = "task-1356503" [ 776.801315] env[61906]: _type = "Task" [ 776.801315] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.814574] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356503, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.830076] env[61906]: DEBUG nova.network.neutron [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.876756] env[61906]: DEBUG nova.network.neutron [req-d93e9082-68bd-4515-932c-21e7bcc85c81 req-f29b34eb-1479-47e6-8234-2f8a57316c71 service nova] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.933920] env[61906]: DEBUG nova.network.neutron [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.963783] env[61906]: DEBUG nova.network.neutron [req-d93e9082-68bd-4515-932c-21e7bcc85c81 req-f29b34eb-1479-47e6-8234-2f8a57316c71 service nova] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.165768] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff70a51-3c6c-43e6-b528-54fbfca2fe32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.173111] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc0a06f-8b15-4ffa-a398-11761ebd495d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.202229] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2585cbde-2105-418c-b5c5-91903d7e1fc5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.209672] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08faa893-5d2c-48b7-a115-76e918e01770 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.222537] env[61906]: DEBUG nova.compute.provider_tree [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.313970] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356503, 'name': ReconfigVM_Task, 'duration_secs': 0.287556} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.314277] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Reconfigured VM instance instance-0000003a to attach disk [datastore2] a6e54432-a58e-49ba-a5c8-e6188f2ea4ff/a6e54432-a58e-49ba-a5c8-e6188f2ea4ff.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 777.314901] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17733234-b6ae-4be0-a609-5240a673cd36 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.320986] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 777.320986] env[61906]: value = "task-1356504" [ 777.320986] env[61906]: _type = "Task" [ 777.320986] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.329047] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.329306] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.334149] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356504, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.439487] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Releasing lock "refresh_cache-a8890726-6c64-4e09-a9a4-0726aa2e8c31" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.439741] env[61906]: DEBUG nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 777.439938] env[61906]: DEBUG nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 777.440178] env[61906]: DEBUG nova.network.neutron [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 777.455525] env[61906]: DEBUG nova.network.neutron [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.469447] env[61906]: DEBUG oslo_concurrency.lockutils [req-d93e9082-68bd-4515-932c-21e7bcc85c81 req-f29b34eb-1479-47e6-8234-2f8a57316c71 service nova] Releasing lock "refresh_cache-b1429121-3c7e-4982-9059-5eb98a7f92df" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.469825] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "refresh_cache-b1429121-3c7e-4982-9059-5eb98a7f92df" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.469999] env[61906]: DEBUG nova.network.neutron [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.725991] env[61906]: DEBUG nova.scheduler.client.report [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.832013] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356504, 'name': Rename_Task, 'duration_secs': 0.167678} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.832013] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 777.832013] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1270ebe-e04a-4e1d-96eb-e6070975829a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.837317] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 777.837317] env[61906]: value = "task-1356505" [ 777.837317] env[61906]: _type = "Task" [ 777.837317] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.844569] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356505, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.959047] env[61906]: DEBUG nova.network.neutron [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.991816] env[61906]: DEBUG nova.network.neutron [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.065042] env[61906]: DEBUG nova.network.neutron [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.230993] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.929s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.231665] env[61906]: ERROR nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cd611107-eb27-4ac8-b1f6-9a29c1516ac7, please check neutron logs for more information. [ 778.231665] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Traceback (most recent call last): [ 778.231665] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 778.231665] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] self.driver.spawn(context, instance, image_meta, [ 778.231665] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 778.231665] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 778.231665] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 778.231665] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] vm_ref = self.build_virtual_machine(instance, [ 778.231665] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 778.231665] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] vif_infos = vmwarevif.get_vif_info(self._session, [ 778.231665] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 778.231940] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] for vif in network_info: [ 778.231940] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 778.231940] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] return self._sync_wrapper(fn, *args, **kwargs) [ 778.231940] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 778.231940] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] self.wait() [ 778.231940] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 778.231940] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] self[:] = self._gt.wait() [ 778.231940] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 778.231940] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] return self._exit_event.wait() [ 778.231940] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 778.231940] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] result = hub.switch() [ 778.231940] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 778.231940] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] return self.greenlet.switch() [ 778.232239] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 778.232239] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] result = function(*args, **kwargs) [ 778.232239] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 778.232239] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] return func(*args, **kwargs) [ 778.232239] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 778.232239] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] raise e [ 778.232239] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 778.232239] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] nwinfo = self.network_api.allocate_for_instance( [ 778.232239] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 778.232239] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] created_port_ids = self._update_ports_for_instance( [ 778.232239] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 778.232239] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] with excutils.save_and_reraise_exception(): [ 778.232239] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.232537] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] self.force_reraise() [ 778.232537] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.232537] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] raise self.value [ 778.232537] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 778.232537] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] updated_port = self._update_port( [ 778.232537] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.232537] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] _ensure_no_port_binding_failure(port) [ 778.232537] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.232537] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] raise exception.PortBindingFailed(port_id=port['id']) [ 778.232537] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] nova.exception.PortBindingFailed: Binding failed for port cd611107-eb27-4ac8-b1f6-9a29c1516ac7, please check neutron logs for more information. [ 778.232537] env[61906]: ERROR nova.compute.manager [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] [ 778.232779] env[61906]: DEBUG nova.compute.utils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Binding failed for port cd611107-eb27-4ac8-b1f6-9a29c1516ac7, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 778.233620] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.661s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.235559] env[61906]: INFO nova.compute.claims [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 778.238750] env[61906]: DEBUG nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Build of instance ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb was re-scheduled: Binding failed for port cd611107-eb27-4ac8-b1f6-9a29c1516ac7, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 778.238750] env[61906]: DEBUG nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 778.238967] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Acquiring lock "refresh_cache-ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.239143] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Acquired lock "refresh_cache-ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.239302] env[61906]: DEBUG nova.network.neutron [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 778.347892] env[61906]: DEBUG oslo_vmware.api [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356505, 'name': PowerOnVM_Task, 'duration_secs': 0.405525} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.348303] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 778.348558] env[61906]: INFO nova.compute.manager [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Took 4.33 seconds to spawn the instance on the hypervisor. [ 778.348782] env[61906]: DEBUG nova.compute.manager [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 778.349597] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337ad672-c71d-45bf-9ea4-a9251989f172 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.370977] env[61906]: DEBUG nova.compute.manager [req-d0a8c634-aeeb-45a8-bbcd-d785e4811aa8 req-4b9af104-c413-4962-bbc0-72497341ffe0 service nova] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Received event network-vif-deleted-dd3c4c33-8c85-4958-9053-440584018c37 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.461820] env[61906]: INFO nova.compute.manager [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] [instance: a8890726-6c64-4e09-a9a4-0726aa2e8c31] Took 1.02 seconds to deallocate network for instance. [ 778.566953] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "refresh_cache-b1429121-3c7e-4982-9059-5eb98a7f92df" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.567418] env[61906]: DEBUG nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 778.567610] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 778.567888] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1130e53-cafe-4032-81e9-fee2f004557a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.576579] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c31c7494-c910-4a5a-9eeb-38a0ec7047ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.598737] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b1429121-3c7e-4982-9059-5eb98a7f92df could not be found. [ 778.598947] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 778.599143] env[61906]: INFO nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Took 0.03 seconds to destroy the instance on the hypervisor. [ 778.599381] env[61906]: DEBUG oslo.service.loopingcall [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.599596] env[61906]: DEBUG nova.compute.manager [-] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.599689] env[61906]: DEBUG nova.network.neutron [-] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 778.613026] env[61906]: DEBUG nova.network.neutron [-] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.765388] env[61906]: DEBUG nova.network.neutron [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.848336] env[61906]: DEBUG nova.network.neutron [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.869172] env[61906]: INFO nova.compute.manager [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Took 34.04 seconds to build instance. [ 779.115009] env[61906]: DEBUG nova.network.neutron [-] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.352015] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Releasing lock "refresh_cache-ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.352015] env[61906]: DEBUG nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 779.352015] env[61906]: DEBUG nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 779.352015] env[61906]: DEBUG nova.network.neutron [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 779.365735] env[61906]: DEBUG nova.network.neutron [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.370347] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7e953852-a801-40e2-8882-cdd64df64892 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Lock "a6e54432-a58e-49ba-a5c8-e6188f2ea4ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.374s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.480022] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d21a0c4-de6c-4854-b78f-e642a5defbd5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.488074] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02dd4518-9498-449a-9e6e-615342fd6fc1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.492304] env[61906]: INFO nova.scheduler.client.report [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Deleted allocations for instance a8890726-6c64-4e09-a9a4-0726aa2e8c31 [ 779.530135] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7099b689-1607-424f-8c26-329aee684680 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.537647] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ff29261-468e-4a28-beb1-3f6d2c06f8cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.551303] env[61906]: DEBUG nova.compute.provider_tree [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.613063] env[61906]: INFO nova.compute.manager [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Rebuilding instance [ 779.616658] env[61906]: INFO nova.compute.manager [-] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Took 1.02 seconds to deallocate network for instance. [ 779.619505] env[61906]: DEBUG nova.compute.claims [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 779.619692] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.653196] env[61906]: DEBUG nova.compute.manager [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 779.654061] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a3ba8a-00cc-4176-9a96-620c18f29930 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.869090] env[61906]: DEBUG nova.network.neutron [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.875105] env[61906]: DEBUG nova.compute.manager [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 780.000435] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dbd715bd-887a-4f26-9386-31a6aa27c93a tempest-ServerPasswordTestJSON-183085716 tempest-ServerPasswordTestJSON-183085716-project-member] Lock "a8890726-6c64-4e09-a9a4-0726aa2e8c31" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.523s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.054024] env[61906]: DEBUG nova.scheduler.client.report [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.165205] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 780.165820] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a49abdf0-b837-4265-bbc0-bd539f09c76c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.173386] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 780.173386] env[61906]: value = "task-1356506" [ 780.173386] env[61906]: _type = "Task" [ 780.173386] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.183560] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356506, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.371729] env[61906]: INFO nova.compute.manager [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] [instance: ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb] Took 1.02 seconds to deallocate network for instance. [ 780.394083] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.506102] env[61906]: DEBUG nova.compute.manager [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 780.559276] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.325s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.559634] env[61906]: DEBUG nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 780.562859] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.649s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.564532] env[61906]: INFO nova.compute.claims [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 780.686702] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356506, 'name': PowerOffVM_Task, 'duration_secs': 0.172376} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.687010] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 780.687234] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.687975] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58218668-3701-477a-8a39-1d24e92a6599 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.694701] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 780.694974] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7536d86c-edfd-463b-96ae-edb25aa80f1b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.723819] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 780.724085] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 780.724277] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Deleting the datastore file [datastore2] a6e54432-a58e-49ba-a5c8-e6188f2ea4ff {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 780.724553] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-00cd1218-e321-443b-a283-ee8799642446 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.731021] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 780.731021] env[61906]: value = "task-1356508" [ 780.731021] env[61906]: _type = "Task" [ 780.731021] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.738897] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356508, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.029326] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.069498] env[61906]: DEBUG nova.compute.utils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 781.075619] env[61906]: DEBUG nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 781.075619] env[61906]: DEBUG nova.network.neutron [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 781.126099] env[61906]: DEBUG nova.policy [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '70a294a034cf41fa93084149243c880e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '256a7dcc6f1549b58e81c41509c0500e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 781.246072] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356508, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106451} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.246493] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 781.246937] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 781.247170] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 781.427979] env[61906]: INFO nova.scheduler.client.report [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Deleted allocations for instance ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb [ 781.495304] env[61906]: DEBUG nova.network.neutron [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Successfully created port: 52f119bb-d26b-495b-945b-9350e74ffb03 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 781.575899] env[61906]: DEBUG nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 781.828145] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6682ad8e-1020-48a1-be9a-f4a08703a430 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.840722] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03112135-5287-4d8c-bf1a-f8534e42b4d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.886128] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ac8875-0d77-489f-80be-003f106e5dbb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.892759] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7769cc-d1a1-4243-8843-2536f39554df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.916152] env[61906]: DEBUG nova.compute.provider_tree [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.940299] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2bafe87-0537-4e9c-8415-5fe1d37718c5 tempest-ServerMetadataNegativeTestJSON-1211414413 tempest-ServerMetadataNegativeTestJSON-1211414413-project-member] Lock "ea2154b0-42ea-4b3b-b874-3a57cbe6bcbb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.448s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.288556] env[61906]: DEBUG nova.virt.hardware [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 782.288808] env[61906]: DEBUG nova.virt.hardware [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 782.288968] env[61906]: DEBUG nova.virt.hardware [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 782.289164] env[61906]: DEBUG nova.virt.hardware [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 782.289312] env[61906]: DEBUG nova.virt.hardware [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 782.289459] env[61906]: DEBUG nova.virt.hardware [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 782.289668] env[61906]: DEBUG nova.virt.hardware [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 782.289827] env[61906]: DEBUG nova.virt.hardware [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 782.289996] env[61906]: DEBUG nova.virt.hardware [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 782.290174] env[61906]: DEBUG nova.virt.hardware [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 782.290916] env[61906]: DEBUG nova.virt.hardware [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.291855] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5aa6401-6edb-48dc-8f21-f35fe3b95fae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.299800] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef02139-494f-456a-9dee-68ee0b4aac52 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.315156] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.321957] env[61906]: DEBUG oslo.service.loopingcall [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.322198] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 782.322407] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a92f46d-a1c4-4548-8100-d9bdd65e4d7e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.337452] env[61906]: DEBUG nova.compute.manager [req-8a4663b9-60ea-4c9b-82d7-a21e4cf1edfb req-023287d0-c1d8-436a-b524-b38df8626b2d service nova] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Received event network-changed-52f119bb-d26b-495b-945b-9350e74ffb03 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.337646] env[61906]: DEBUG nova.compute.manager [req-8a4663b9-60ea-4c9b-82d7-a21e4cf1edfb req-023287d0-c1d8-436a-b524-b38df8626b2d service nova] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Refreshing instance network info cache due to event network-changed-52f119bb-d26b-495b-945b-9350e74ffb03. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 782.337863] env[61906]: DEBUG oslo_concurrency.lockutils [req-8a4663b9-60ea-4c9b-82d7-a21e4cf1edfb req-023287d0-c1d8-436a-b524-b38df8626b2d service nova] Acquiring lock "refresh_cache-c3cb1a2d-e905-497c-a4de-8c6579c37156" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.338012] env[61906]: DEBUG oslo_concurrency.lockutils [req-8a4663b9-60ea-4c9b-82d7-a21e4cf1edfb req-023287d0-c1d8-436a-b524-b38df8626b2d service nova] Acquired lock "refresh_cache-c3cb1a2d-e905-497c-a4de-8c6579c37156" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.338179] env[61906]: DEBUG nova.network.neutron [req-8a4663b9-60ea-4c9b-82d7-a21e4cf1edfb req-023287d0-c1d8-436a-b524-b38df8626b2d service nova] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Refreshing network info cache for port 52f119bb-d26b-495b-945b-9350e74ffb03 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 782.345951] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.345951] env[61906]: value = "task-1356509" [ 782.345951] env[61906]: _type = "Task" [ 782.345951] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.355806] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356509, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.419387] env[61906]: DEBUG nova.scheduler.client.report [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.443389] env[61906]: DEBUG nova.compute.manager [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 782.541059] env[61906]: ERROR nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 52f119bb-d26b-495b-945b-9350e74ffb03, please check neutron logs for more information. [ 782.541059] env[61906]: ERROR nova.compute.manager Traceback (most recent call last): [ 782.541059] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.541059] env[61906]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 782.541059] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 782.541059] env[61906]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 782.541059] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 782.541059] env[61906]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 782.541059] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.541059] env[61906]: ERROR nova.compute.manager self.force_reraise() [ 782.541059] env[61906]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.541059] env[61906]: ERROR nova.compute.manager raise self.value [ 782.541059] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 782.541059] env[61906]: ERROR nova.compute.manager updated_port = self._update_port( [ 782.541059] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.541059] env[61906]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 782.541712] env[61906]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.541712] env[61906]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 782.541712] env[61906]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 52f119bb-d26b-495b-945b-9350e74ffb03, please check neutron logs for more information. [ 782.541712] env[61906]: ERROR nova.compute.manager [ 782.541712] env[61906]: Traceback (most recent call last): [ 782.541712] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 782.541712] env[61906]: listener.cb(fileno) [ 782.541712] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 782.541712] env[61906]: result = function(*args, **kwargs) [ 782.541712] env[61906]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 782.541712] env[61906]: return func(*args, **kwargs) [ 782.541712] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 782.541712] env[61906]: raise e [ 782.541712] env[61906]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.541712] env[61906]: nwinfo = self.network_api.allocate_for_instance( [ 782.541712] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 782.541712] env[61906]: created_port_ids = self._update_ports_for_instance( [ 782.541712] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 782.541712] env[61906]: with excutils.save_and_reraise_exception(): [ 782.541712] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.541712] env[61906]: self.force_reraise() [ 782.541712] env[61906]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.541712] env[61906]: raise self.value [ 782.541712] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 782.541712] env[61906]: updated_port = self._update_port( [ 782.541712] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.541712] env[61906]: _ensure_no_port_binding_failure(port) [ 782.541712] env[61906]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.541712] env[61906]: raise exception.PortBindingFailed(port_id=port['id']) [ 782.542585] env[61906]: nova.exception.PortBindingFailed: Binding failed for port 52f119bb-d26b-495b-945b-9350e74ffb03, please check neutron logs for more information. [ 782.542585] env[61906]: Removing descriptor: 15 [ 782.590916] env[61906]: DEBUG nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 782.611620] env[61906]: DEBUG nova.virt.hardware [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 782.611909] env[61906]: DEBUG nova.virt.hardware [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 782.612085] env[61906]: DEBUG nova.virt.hardware [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 782.612270] env[61906]: DEBUG nova.virt.hardware [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 782.612416] env[61906]: DEBUG nova.virt.hardware [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 782.612635] env[61906]: DEBUG nova.virt.hardware [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 782.612857] env[61906]: DEBUG nova.virt.hardware [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 782.613026] env[61906]: DEBUG nova.virt.hardware [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 782.613197] env[61906]: DEBUG nova.virt.hardware [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 782.613360] env[61906]: DEBUG nova.virt.hardware [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 782.613529] env[61906]: DEBUG nova.virt.hardware [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.614386] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bad1705-1dae-4815-a587-d78199513d3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.624499] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85759d4e-4e68-4f7a-8bdd-d32949b5c6ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.637944] env[61906]: ERROR nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 52f119bb-d26b-495b-945b-9350e74ffb03, please check neutron logs for more information. [ 782.637944] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Traceback (most recent call last): [ 782.637944] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 782.637944] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] yield resources [ 782.637944] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 782.637944] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] self.driver.spawn(context, instance, image_meta, [ 782.637944] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 782.637944] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] self._vmops.spawn(context, instance, image_meta, injected_files, [ 782.637944] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 782.637944] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] vm_ref = self.build_virtual_machine(instance, [ 782.637944] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 782.638410] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] vif_infos = vmwarevif.get_vif_info(self._session, [ 782.638410] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 782.638410] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] for vif in network_info: [ 782.638410] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 782.638410] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] return self._sync_wrapper(fn, *args, **kwargs) [ 782.638410] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 782.638410] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] self.wait() [ 782.638410] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 782.638410] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] self[:] = self._gt.wait() [ 782.638410] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 782.638410] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] return self._exit_event.wait() [ 782.638410] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 782.638410] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] current.throw(*self._exc) [ 782.638764] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 782.638764] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] result = function(*args, **kwargs) [ 782.638764] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 782.638764] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] return func(*args, **kwargs) [ 782.638764] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 782.638764] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] raise e [ 782.638764] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.638764] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] nwinfo = self.network_api.allocate_for_instance( [ 782.638764] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 782.638764] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] created_port_ids = self._update_ports_for_instance( [ 782.638764] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 782.638764] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] with excutils.save_and_reraise_exception(): [ 782.638764] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.639110] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] self.force_reraise() [ 782.639110] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.639110] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] raise self.value [ 782.639110] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 782.639110] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] updated_port = self._update_port( [ 782.639110] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.639110] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] _ensure_no_port_binding_failure(port) [ 782.639110] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.639110] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] raise exception.PortBindingFailed(port_id=port['id']) [ 782.639110] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] nova.exception.PortBindingFailed: Binding failed for port 52f119bb-d26b-495b-945b-9350e74ffb03, please check neutron logs for more information. [ 782.639110] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] [ 782.639110] env[61906]: INFO nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Terminating instance [ 782.640278] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquiring lock "refresh_cache-c3cb1a2d-e905-497c-a4de-8c6579c37156" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.855890] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356509, 'name': CreateVM_Task, 'duration_secs': 0.248791} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.856697] env[61906]: DEBUG nova.network.neutron [req-8a4663b9-60ea-4c9b-82d7-a21e4cf1edfb req-023287d0-c1d8-436a-b524-b38df8626b2d service nova] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.858293] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 782.859183] env[61906]: DEBUG oslo_vmware.service [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0104486d-0f05-474e-9794-d2f7ca1a1c98 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.865259] env[61906]: DEBUG oslo_concurrency.lockutils [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.865424] env[61906]: DEBUG oslo_concurrency.lockutils [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.865812] env[61906]: DEBUG oslo_concurrency.lockutils [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.866069] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1bf4ae6-4fac-41fe-a8b1-235373e2a7a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.876033] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 782.876033] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5271a1ad-83a9-2c86-c0d2-9bcfb60c14ad" [ 782.876033] env[61906]: _type = "Task" [ 782.876033] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.881210] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5271a1ad-83a9-2c86-c0d2-9bcfb60c14ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.927569] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.927831] env[61906]: DEBUG nova.compute.manager [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 782.933862] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.835s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.935418] env[61906]: INFO nova.compute.claims [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.952256] env[61906]: DEBUG nova.network.neutron [req-8a4663b9-60ea-4c9b-82d7-a21e4cf1edfb req-023287d0-c1d8-436a-b524-b38df8626b2d service nova] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.983965] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.389541] env[61906]: DEBUG oslo_concurrency.lockutils [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.389621] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.389837] env[61906]: DEBUG oslo_concurrency.lockutils [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.390540] env[61906]: DEBUG oslo_concurrency.lockutils [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.390540] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.390693] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9baaaecb-cb53-4c42-9bfa-178441574555 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.400352] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.400547] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.401409] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8f10af5-776b-4e45-9c37-4ed057710a0f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.411121] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c81a96f-b75c-4c07-983d-891a1cffb25b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.417549] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 783.417549] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528b0718-9a38-82eb-871a-55c3f4a245f6" [ 783.417549] env[61906]: _type = "Task" [ 783.417549] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.425485] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528b0718-9a38-82eb-871a-55c3f4a245f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.442039] env[61906]: DEBUG nova.compute.utils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 783.445485] env[61906]: DEBUG nova.compute.manager [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 783.445747] env[61906]: DEBUG nova.network.neutron [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 783.454717] env[61906]: DEBUG oslo_concurrency.lockutils [req-8a4663b9-60ea-4c9b-82d7-a21e4cf1edfb req-023287d0-c1d8-436a-b524-b38df8626b2d service nova] Releasing lock "refresh_cache-c3cb1a2d-e905-497c-a4de-8c6579c37156" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.455135] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquired lock "refresh_cache-c3cb1a2d-e905-497c-a4de-8c6579c37156" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.455323] env[61906]: DEBUG nova.network.neutron [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.520938] env[61906]: DEBUG nova.policy [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a78f9d00d14e4de19318d6dfa1ab47ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78d00c69a4e34417b349af62d0d69bc8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 783.669781] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "a5f780a2-0cb1-4da8-8276-82b9653bf6a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.669781] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "a5f780a2-0cb1-4da8-8276-82b9653bf6a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.817538] env[61906]: DEBUG nova.network.neutron [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Successfully created port: 69416cc5-266f-493b-bc21-5288dbcec92e {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 783.934088] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Preparing fetch location {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 783.934696] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Creating directory with path [datastore1] vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.939873] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-190a3295-1698-407b-8211-375d844c9fb7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.946171] env[61906]: DEBUG nova.compute.manager [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 783.961906] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Created directory with path [datastore1] vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.961906] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Fetch image to [datastore1] vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 783.961906] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Downloading image file data 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 to [datastore1] vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk on the data store datastore1 {{(pid=61906) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 783.962148] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071add6b-4535-41c2-8920-755ea078962e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.970607] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e06bed-d536-4488-9d48-43bbecb61e7f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.984031] env[61906]: DEBUG nova.network.neutron [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.986437] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf9a6fe-b05b-4cb2-b91e-95ac4f4f1600 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.996515] env[61906]: DEBUG oslo_concurrency.lockutils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "31d0f640-47f4-4b72-b9fe-6ef4c76c6b44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.997014] env[61906]: DEBUG oslo_concurrency.lockutils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "31d0f640-47f4-4b72-b9fe-6ef4c76c6b44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.034595] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ffc4cf-bf51-4dda-b116-5c76b1cdfaac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.041619] env[61906]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-eea73387-5b3a-4673-82f5-563bbb80d935 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.072556] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Downloading image file data 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 to the data store datastore1 {{(pid=61906) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 784.154419] env[61906]: DEBUG oslo_vmware.rw_handles [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61906) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 784.156517] env[61906]: DEBUG nova.network.neutron [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.363128] env[61906]: DEBUG nova.compute.manager [req-9a10d881-1937-45dd-95fc-e366c02cdbf4 req-6a6e10df-c42c-4de4-b3f2-dadab8197308 service nova] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Received event network-vif-deleted-52f119bb-d26b-495b-945b-9350e74ffb03 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.400515] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4918ac-226d-417f-acc3-17b51b28b17a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.408051] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81055d95-d06a-4f3f-b7ab-d6df857eed56 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.444645] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99cc332b-5a66-43d2-9d67-816dc3764d78 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.458152] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-324abec9-fed4-4a09-98ec-52fadfd9fe5b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.472883] env[61906]: DEBUG nova.compute.provider_tree [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.664050] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Releasing lock "refresh_cache-c3cb1a2d-e905-497c-a4de-8c6579c37156" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.664210] env[61906]: DEBUG nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 784.664772] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 784.664772] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2dab2626-2ef4-4e22-98dc-f09f57b4a365 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.674850] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a18527-0ba1-4d13-8165-6323f2ee1a30 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.701514] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c3cb1a2d-e905-497c-a4de-8c6579c37156 could not be found. [ 784.701804] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 784.701995] env[61906]: INFO nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Took 0.04 seconds to destroy the instance on the hypervisor. [ 784.702369] env[61906]: DEBUG oslo.service.loopingcall [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.703088] env[61906]: DEBUG nova.compute.manager [-] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.703189] env[61906]: DEBUG nova.network.neutron [-] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.720199] env[61906]: DEBUG nova.network.neutron [-] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.777276] env[61906]: DEBUG oslo_vmware.rw_handles [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Completed reading data from the image iterator. {{(pid=61906) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 784.777514] env[61906]: DEBUG oslo_vmware.rw_handles [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 784.921401] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Downloaded image file data 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 to vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk on the data store datastore1 {{(pid=61906) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 784.923686] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Caching image {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 784.923983] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Copying Virtual Disk [datastore1] vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk to [datastore1] vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 784.924271] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3dd704e7-bf58-48f3-93c3-b907e5b4c84b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.931726] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 784.931726] env[61906]: value = "task-1356510" [ 784.931726] env[61906]: _type = "Task" [ 784.931726] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.941863] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356510, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.963212] env[61906]: DEBUG nova.compute.manager [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 784.976123] env[61906]: DEBUG nova.scheduler.client.report [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.989582] env[61906]: DEBUG nova.virt.hardware [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 784.989820] env[61906]: DEBUG nova.virt.hardware [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 784.989975] env[61906]: DEBUG nova.virt.hardware [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 784.990168] env[61906]: DEBUG nova.virt.hardware [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 784.990310] env[61906]: DEBUG nova.virt.hardware [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 784.990453] env[61906]: DEBUG nova.virt.hardware [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 784.990650] env[61906]: DEBUG nova.virt.hardware [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 784.990805] env[61906]: DEBUG nova.virt.hardware [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 784.990964] env[61906]: DEBUG nova.virt.hardware [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 784.991133] env[61906]: DEBUG nova.virt.hardware [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 784.991305] env[61906]: DEBUG nova.virt.hardware [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 784.992869] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e16ba1-93b6-43e7-b9a2-4ace4bf41c1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.001190] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-816bbc6a-333c-4fbe-9f3a-36bc42b301cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.223294] env[61906]: DEBUG nova.network.neutron [-] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.441932] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356510, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.483087] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.483705] env[61906]: DEBUG nova.compute.manager [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.488403] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.357s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.499075] env[61906]: DEBUG nova.network.neutron [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Successfully updated port: 69416cc5-266f-493b-bc21-5288dbcec92e {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 785.727600] env[61906]: INFO nova.compute.manager [-] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Took 1.02 seconds to deallocate network for instance. [ 785.732684] env[61906]: DEBUG nova.compute.claims [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Aborting claim: {{(pid=61906) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 785.732913] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.942889] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356510, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.655692} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.943297] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Copied Virtual Disk [datastore1] vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk to [datastore1] vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 785.943557] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Deleting the datastore file [datastore1] vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/tmp-sparse.vmdk {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 785.943827] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-894f0b0f-a679-47ac-aaa5-cbc9ac0ed663 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.950791] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 785.950791] env[61906]: value = "task-1356511" [ 785.950791] env[61906]: _type = "Task" [ 785.950791] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.958205] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356511, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.997470] env[61906]: DEBUG nova.compute.utils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.998886] env[61906]: DEBUG nova.compute.manager [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 785.999076] env[61906]: DEBUG nova.network.neutron [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 786.003963] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Acquiring lock "refresh_cache-77791983-cef1-4f07-9b62-d52d335b08c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.004130] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Acquired lock "refresh_cache-77791983-cef1-4f07-9b62-d52d335b08c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.004474] env[61906]: DEBUG nova.network.neutron [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 786.042682] env[61906]: DEBUG nova.policy [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '217c5f7019504ef8a4d4b11fdbb2a343', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d5318e962dd4c17af474c9d0786a414', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 786.263734] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61bdfae5-6655-4378-bc0f-76ac8c655b2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.273210] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053168fe-ba6e-4d08-83bf-bf7c295ce499 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.306676] env[61906]: DEBUG nova.network.neutron [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Successfully created port: f38d9ad1-e3be-4227-981d-b2b14c6cb912 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.309176] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897aecd4-7273-4a8c-b0f8-28532fd8dc00 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.316753] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d657ade2-0280-4c16-a1e0-59ffd2398f59 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.330413] env[61906]: DEBUG nova.compute.provider_tree [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.461828] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356511, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024132} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.462141] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 786.462522] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Moving file from [datastore1] vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 to [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1. {{(pid=61906) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 786.462841] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-d8fe14f3-b6fc-4968-8a72-bafc9426eb16 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.467540] env[61906]: DEBUG nova.compute.manager [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Received event network-vif-plugged-69416cc5-266f-493b-bc21-5288dbcec92e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.467746] env[61906]: DEBUG oslo_concurrency.lockutils [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] Acquiring lock "77791983-cef1-4f07-9b62-d52d335b08c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.467949] env[61906]: DEBUG oslo_concurrency.lockutils [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] Lock "77791983-cef1-4f07-9b62-d52d335b08c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.468307] env[61906]: DEBUG oslo_concurrency.lockutils [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] Lock "77791983-cef1-4f07-9b62-d52d335b08c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.468483] env[61906]: DEBUG nova.compute.manager [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] No waiting events found dispatching network-vif-plugged-69416cc5-266f-493b-bc21-5288dbcec92e {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 786.468650] env[61906]: WARNING nova.compute.manager [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Received unexpected event network-vif-plugged-69416cc5-266f-493b-bc21-5288dbcec92e for instance with vm_state building and task_state spawning. [ 786.468821] env[61906]: DEBUG nova.compute.manager [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Received event network-changed-69416cc5-266f-493b-bc21-5288dbcec92e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.469085] env[61906]: DEBUG nova.compute.manager [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Refreshing instance network info cache due to event network-changed-69416cc5-266f-493b-bc21-5288dbcec92e. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 786.469147] env[61906]: DEBUG oslo_concurrency.lockutils [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] Acquiring lock "refresh_cache-77791983-cef1-4f07-9b62-d52d335b08c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.473939] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 786.473939] env[61906]: value = "task-1356512" [ 786.473939] env[61906]: _type = "Task" [ 786.473939] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.482697] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356512, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.501905] env[61906]: DEBUG nova.compute.manager [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 786.579303] env[61906]: DEBUG nova.network.neutron [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.834389] env[61906]: DEBUG nova.scheduler.client.report [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 786.903359] env[61906]: DEBUG nova.network.neutron [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Updating instance_info_cache with network_info: [{"id": "69416cc5-266f-493b-bc21-5288dbcec92e", "address": "fa:16:3e:10:b7:52", "network": {"id": "b8f5142a-642a-42be-abab-6f905e38b0a4", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1549034972-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78d00c69a4e34417b349af62d0d69bc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69416cc5-26", "ovs_interfaceid": "69416cc5-266f-493b-bc21-5288dbcec92e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.983756] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356512, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.026151} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.984034] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] File moved {{(pid=61906) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 786.984281] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Cleaning up location [datastore1] vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 786.984460] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Deleting the datastore file [datastore1] vmware_temp/16718872-cb6c-4958-b061-4bab64d65a80 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 786.985083] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c757fc8d-59e8-45d0-9ca7-569977458633 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.991823] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 786.991823] env[61906]: value = "task-1356513" [ 786.991823] env[61906]: _type = "Task" [ 786.991823] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.999461] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356513, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.342881] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.854s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.343667] env[61906]: ERROR nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b3dbfa76-7b63-45cb-bdbb-6045d381ed93, please check neutron logs for more information. [ 787.343667] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Traceback (most recent call last): [ 787.343667] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 787.343667] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] self.driver.spawn(context, instance, image_meta, [ 787.343667] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 787.343667] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 787.343667] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 787.343667] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] vm_ref = self.build_virtual_machine(instance, [ 787.343667] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 787.343667] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] vif_infos = vmwarevif.get_vif_info(self._session, [ 787.343667] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] for vif in network_info: [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] return self._sync_wrapper(fn, *args, **kwargs) [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] self.wait() [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] self[:] = self._gt.wait() [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] return self._exit_event.wait() [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] current.throw(*self._exc) [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] result = function(*args, **kwargs) [ 787.344080] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 787.344448] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] return func(*args, **kwargs) [ 787.344448] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 787.344448] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] raise e [ 787.344448] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.344448] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] nwinfo = self.network_api.allocate_for_instance( [ 787.344448] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.344448] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] created_port_ids = self._update_ports_for_instance( [ 787.344448] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.344448] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] with excutils.save_and_reraise_exception(): [ 787.344448] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.344448] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] self.force_reraise() [ 787.344448] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.344448] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] raise self.value [ 787.344848] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.344848] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] updated_port = self._update_port( [ 787.344848] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.344848] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] _ensure_no_port_binding_failure(port) [ 787.344848] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.344848] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] raise exception.PortBindingFailed(port_id=port['id']) [ 787.344848] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] nova.exception.PortBindingFailed: Binding failed for port b3dbfa76-7b63-45cb-bdbb-6045d381ed93, please check neutron logs for more information. [ 787.344848] env[61906]: ERROR nova.compute.manager [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] [ 787.344848] env[61906]: DEBUG nova.compute.utils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Binding failed for port b3dbfa76-7b63-45cb-bdbb-6045d381ed93, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 787.345629] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.362s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.347259] env[61906]: INFO nova.compute.claims [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.349667] env[61906]: DEBUG nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Build of instance 45621a25-cd9c-4931-899c-647da0b50ec7 was re-scheduled: Binding failed for port b3dbfa76-7b63-45cb-bdbb-6045d381ed93, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 787.350105] env[61906]: DEBUG nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 787.350335] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Acquiring lock "refresh_cache-45621a25-cd9c-4931-899c-647da0b50ec7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.350480] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Acquired lock "refresh_cache-45621a25-cd9c-4931-899c-647da0b50ec7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.350670] env[61906]: DEBUG nova.network.neutron [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 787.406202] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Releasing lock "refresh_cache-77791983-cef1-4f07-9b62-d52d335b08c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.406527] env[61906]: DEBUG nova.compute.manager [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Instance network_info: |[{"id": "69416cc5-266f-493b-bc21-5288dbcec92e", "address": "fa:16:3e:10:b7:52", "network": {"id": "b8f5142a-642a-42be-abab-6f905e38b0a4", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1549034972-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78d00c69a4e34417b349af62d0d69bc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69416cc5-26", "ovs_interfaceid": "69416cc5-266f-493b-bc21-5288dbcec92e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 787.406833] env[61906]: DEBUG oslo_concurrency.lockutils [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] Acquired lock "refresh_cache-77791983-cef1-4f07-9b62-d52d335b08c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.407010] env[61906]: DEBUG nova.network.neutron [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Refreshing network info cache for port 69416cc5-266f-493b-bc21-5288dbcec92e {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 787.408233] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:b7:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1430a695-49fb-4905-bc38-db9b869a1a9d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '69416cc5-266f-493b-bc21-5288dbcec92e', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.417638] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Creating folder: Project (78d00c69a4e34417b349af62d0d69bc8). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 787.418610] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e614badc-e5d1-4860-aac6-d74a69a8d336 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.429556] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Created folder: Project (78d00c69a4e34417b349af62d0d69bc8) in parent group-v288914. [ 787.429747] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Creating folder: Instances. Parent ref: group-v288938. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 787.429980] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e449c9cb-9755-4ede-a4b6-c3321f70e45d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.439616] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Created folder: Instances in parent group-v288938. [ 787.439988] env[61906]: DEBUG oslo.service.loopingcall [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.440101] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 787.440333] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-47f2a0f2-f94b-4497-8c73-a480c7b11154 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.459495] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 787.459495] env[61906]: value = "task-1356516" [ 787.459495] env[61906]: _type = "Task" [ 787.459495] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.467059] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356516, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.502596] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356513, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026245} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.502986] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.503693] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39d4bb10-1e92-437d-afc0-6768f2dc86d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.508849] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 787.508849] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ccf217-6bf5-6f3d-c1b7-1df1681b7583" [ 787.508849] env[61906]: _type = "Task" [ 787.508849] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.517780] env[61906]: DEBUG nova.compute.manager [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.527757] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ccf217-6bf5-6f3d-c1b7-1df1681b7583, 'name': SearchDatastore_Task, 'duration_secs': 0.009696} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.528037] env[61906]: DEBUG oslo_concurrency.lockutils [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.528315] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] a6e54432-a58e-49ba-a5c8-e6188f2ea4ff/a6e54432-a58e-49ba-a5c8-e6188f2ea4ff.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 787.528581] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e6bb10f8-cb20-4687-ba35-6e31e9004589 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.535250] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 787.535250] env[61906]: value = "task-1356517" [ 787.535250] env[61906]: _type = "Task" [ 787.535250] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.545218] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356517, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.552460] env[61906]: DEBUG nova.virt.hardware [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.552725] env[61906]: DEBUG nova.virt.hardware [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.553169] env[61906]: DEBUG nova.virt.hardware [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.553169] env[61906]: DEBUG nova.virt.hardware [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.553257] env[61906]: DEBUG nova.virt.hardware [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.553766] env[61906]: DEBUG nova.virt.hardware [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.553766] env[61906]: DEBUG nova.virt.hardware [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.553766] env[61906]: DEBUG nova.virt.hardware [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.553885] env[61906]: DEBUG nova.virt.hardware [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.553998] env[61906]: DEBUG nova.virt.hardware [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.556935] env[61906]: DEBUG nova.virt.hardware [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.557175] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd0cb34-e1de-49cb-97de-99a5013f6569 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.565548] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f58c7f89-74d2-4e6e-8fad-7fcb759d1178 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.903282] env[61906]: DEBUG nova.network.neutron [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.974380] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356516, 'name': CreateVM_Task, 'duration_secs': 0.347211} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.974868] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 787.988146] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.988146] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.988146] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 787.988146] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0d40faa-df3b-4b2f-9d3b-732e324a794a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.992753] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Waiting for the task: (returnval){ [ 787.992753] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c7247a-5099-207b-0383-86ac54eb2a6e" [ 787.992753] env[61906]: _type = "Task" [ 787.992753] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.004461] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c7247a-5099-207b-0383-86ac54eb2a6e, 'name': SearchDatastore_Task, 'duration_secs': 0.008184} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.008175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.008465] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 788.008701] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.008849] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.009092] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.009317] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1850f8d-d0ae-4f6a-ac87-75a0724049f4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.019019] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.019019] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 788.019019] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf86e32b-82c1-4458-a375-eb25ecd99137 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.024504] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Waiting for the task: (returnval){ [ 788.024504] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ff4141-997c-1321-f296-77b2390fa9ff" [ 788.024504] env[61906]: _type = "Task" [ 788.024504] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.030854] env[61906]: DEBUG nova.network.neutron [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Successfully updated port: f38d9ad1-e3be-4227-981d-b2b14c6cb912 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 788.035123] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ff4141-997c-1321-f296-77b2390fa9ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.045166] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356517, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436793} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.045436] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] a6e54432-a58e-49ba-a5c8-e6188f2ea4ff/a6e54432-a58e-49ba-a5c8-e6188f2ea4ff.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 788.045658] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 788.045907] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-255cbfbe-d013-46d4-b570-1b985ccda675 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.052527] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 788.052527] env[61906]: value = "task-1356518" [ 788.052527] env[61906]: _type = "Task" [ 788.052527] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.065952] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356518, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.162504] env[61906]: DEBUG nova.network.neutron [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.393415] env[61906]: DEBUG nova.network.neutron [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Updated VIF entry in instance network info cache for port 69416cc5-266f-493b-bc21-5288dbcec92e. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 788.393415] env[61906]: DEBUG nova.network.neutron [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Updating instance_info_cache with network_info: [{"id": "69416cc5-266f-493b-bc21-5288dbcec92e", "address": "fa:16:3e:10:b7:52", "network": {"id": "b8f5142a-642a-42be-abab-6f905e38b0a4", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1549034972-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "78d00c69a4e34417b349af62d0d69bc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1430a695-49fb-4905-bc38-db9b869a1a9d", "external-id": "nsx-vlan-transportzone-297", "segmentation_id": 297, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69416cc5-26", "ovs_interfaceid": "69416cc5-266f-493b-bc21-5288dbcec92e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.498347] env[61906]: DEBUG nova.compute.manager [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Received event network-vif-plugged-f38d9ad1-e3be-4227-981d-b2b14c6cb912 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.498347] env[61906]: DEBUG oslo_concurrency.lockutils [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] Acquiring lock "7943ab1a-82df-48dd-874c-15dc015db51e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.498851] env[61906]: DEBUG oslo_concurrency.lockutils [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] Lock "7943ab1a-82df-48dd-874c-15dc015db51e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.498851] env[61906]: DEBUG oslo_concurrency.lockutils [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] Lock "7943ab1a-82df-48dd-874c-15dc015db51e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.498851] env[61906]: DEBUG nova.compute.manager [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] No waiting events found dispatching network-vif-plugged-f38d9ad1-e3be-4227-981d-b2b14c6cb912 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 788.498851] env[61906]: WARNING nova.compute.manager [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Received unexpected event network-vif-plugged-f38d9ad1-e3be-4227-981d-b2b14c6cb912 for instance with vm_state building and task_state spawning. [ 788.499092] env[61906]: DEBUG nova.compute.manager [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Received event network-changed-f38d9ad1-e3be-4227-981d-b2b14c6cb912 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.500587] env[61906]: DEBUG nova.compute.manager [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Refreshing instance network info cache due to event network-changed-f38d9ad1-e3be-4227-981d-b2b14c6cb912. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 788.500587] env[61906]: DEBUG oslo_concurrency.lockutils [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] Acquiring lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.500759] env[61906]: DEBUG oslo_concurrency.lockutils [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] Acquired lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.500833] env[61906]: DEBUG nova.network.neutron [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Refreshing network info cache for port f38d9ad1-e3be-4227-981d-b2b14c6cb912 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 788.539557] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.539788] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ff4141-997c-1321-f296-77b2390fa9ff, 'name': SearchDatastore_Task, 'duration_secs': 0.007991} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.543838] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b93b922c-0482-4552-8575-98fa17cf6dda {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.549534] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Waiting for the task: (returnval){ [ 788.549534] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52139ee4-9372-fd07-828f-070c8d009158" [ 788.549534] env[61906]: _type = "Task" [ 788.549534] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.562452] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52139ee4-9372-fd07-828f-070c8d009158, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.568919] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356518, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062346} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.569210] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 788.570090] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb4c1c6-3861-41f8-8f12-b804dfdf4e3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.590921] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] a6e54432-a58e-49ba-a5c8-e6188f2ea4ff/a6e54432-a58e-49ba-a5c8-e6188f2ea4ff.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 788.592343] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46603728-4e57-4312-8196-82ba295a0483 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.613621] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 788.613621] env[61906]: value = "task-1356519" [ 788.613621] env[61906]: _type = "Task" [ 788.613621] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.621975] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356519, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.668909] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Releasing lock "refresh_cache-45621a25-cd9c-4931-899c-647da0b50ec7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.669188] env[61906]: DEBUG nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 788.669396] env[61906]: DEBUG nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 788.669596] env[61906]: DEBUG nova.network.neutron [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 788.741328] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91963c30-b5e7-4d11-9b0b-d74502df7f40 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.750982] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb19804a-422b-45a8-80ff-9f0f2d8dc3a9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.785009] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8b928c-75b8-4206-9b10-bad63907f705 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.793935] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55527422-4973-4634-837c-4788bcd7a671 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.806302] env[61906]: DEBUG nova.compute.provider_tree [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.840011] env[61906]: DEBUG nova.network.neutron [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.897034] env[61906]: DEBUG oslo_concurrency.lockutils [req-f81f419b-0be0-40de-9904-7156ace4191e req-15036673-2f41-4919-99a1-8a2a2b06a34a service nova] Releasing lock "refresh_cache-77791983-cef1-4f07-9b62-d52d335b08c8" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.061340] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52139ee4-9372-fd07-828f-070c8d009158, 'name': SearchDatastore_Task, 'duration_secs': 0.009754} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.061847] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.062264] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 77791983-cef1-4f07-9b62-d52d335b08c8/77791983-cef1-4f07-9b62-d52d335b08c8.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 789.063029] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9d9ab6e-436c-47a0-8f74-0287de1626d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.068611] env[61906]: DEBUG nova.network.neutron [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.079522] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Waiting for the task: (returnval){ [ 789.079522] env[61906]: value = "task-1356520" [ 789.079522] env[61906]: _type = "Task" [ 789.079522] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.086974] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356520, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.128421] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356519, 'name': ReconfigVM_Task, 'duration_secs': 0.259971} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.130527] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Reconfigured VM instance instance-0000003a to attach disk [datastore1] a6e54432-a58e-49ba-a5c8-e6188f2ea4ff/a6e54432-a58e-49ba-a5c8-e6188f2ea4ff.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.130527] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-499e0d1c-2f7b-44c5-b29b-b1fb662a9d3e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.139226] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 789.139226] env[61906]: value = "task-1356521" [ 789.139226] env[61906]: _type = "Task" [ 789.139226] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.148089] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356521, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.216032] env[61906]: DEBUG nova.network.neutron [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.309452] env[61906]: DEBUG nova.scheduler.client.report [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.343277] env[61906]: DEBUG nova.network.neutron [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.588467] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356520, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450608} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.588467] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 77791983-cef1-4f07-9b62-d52d335b08c8/77791983-cef1-4f07-9b62-d52d335b08c8.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 789.588467] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.588467] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c6a8cccd-8d04-4c3e-bcec-7722313ec51e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.595346] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Waiting for the task: (returnval){ [ 789.595346] env[61906]: value = "task-1356522" [ 789.595346] env[61906]: _type = "Task" [ 789.595346] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.611252] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356522, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.649637] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356521, 'name': Rename_Task, 'duration_secs': 0.144794} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.649847] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 789.650105] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0fd0f670-9709-442c-9259-79a16997c301 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.655588] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 789.655588] env[61906]: value = "task-1356523" [ 789.655588] env[61906]: _type = "Task" [ 789.655588] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.664219] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356523, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.722968] env[61906]: DEBUG oslo_concurrency.lockutils [req-5d89a7eb-1a25-45e0-a1bb-185311531ba0 req-3435b763-efc3-44ae-a8fd-8805da88e879 service nova] Releasing lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.723914] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquired lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.725449] env[61906]: DEBUG nova.network.neutron [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 789.822164] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.822164] env[61906]: DEBUG nova.compute.manager [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 789.822895] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.434s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.849057] env[61906]: INFO nova.compute.manager [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] [instance: 45621a25-cd9c-4931-899c-647da0b50ec7] Took 1.18 seconds to deallocate network for instance. [ 790.107956] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356522, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058548} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.108320] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 790.109150] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b955c406-4ed9-4f97-bd5c-1c2d2cc1047c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.131477] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 77791983-cef1-4f07-9b62-d52d335b08c8/77791983-cef1-4f07-9b62-d52d335b08c8.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 790.131849] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd0ac76a-393d-400d-b19f-7b5a309fac4f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.150376] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Waiting for the task: (returnval){ [ 790.150376] env[61906]: value = "task-1356524" [ 790.150376] env[61906]: _type = "Task" [ 790.150376] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.159245] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356524, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.167830] env[61906]: DEBUG oslo_vmware.api [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356523, 'name': PowerOnVM_Task, 'duration_secs': 0.461129} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.168401] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 790.169078] env[61906]: DEBUG nova.compute.manager [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 790.169903] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b037b166-067b-4d26-9023-9a9104a08097 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.265785] env[61906]: DEBUG nova.network.neutron [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.329128] env[61906]: DEBUG nova.compute.utils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.336431] env[61906]: DEBUG nova.compute.manager [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 790.336431] env[61906]: DEBUG nova.network.neutron [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 790.455447] env[61906]: DEBUG nova.policy [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ded7bad07b5e4f31ba627d4e4dbea0d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '473163e61f064b969326e695af5c54b1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 790.573062] env[61906]: DEBUG nova.network.neutron [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Updating instance_info_cache with network_info: [{"id": "f38d9ad1-e3be-4227-981d-b2b14c6cb912", "address": "fa:16:3e:24:7a:87", "network": {"id": "1844d418-05f6-4457-83a6-08833666041a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1376388956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d5318e962dd4c17af474c9d0786a414", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed8a78a1-87dc-488e-a092-afd1c2a2ddde", "external-id": "nsx-vlan-transportzone-21", "segmentation_id": 21, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf38d9ad1-e3", "ovs_interfaceid": "f38d9ad1-e3be-4227-981d-b2b14c6cb912", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.663319] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356524, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.688985] env[61906]: DEBUG oslo_concurrency.lockutils [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.712824] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81107768-9c78-4466-8f44-1e220cec4985 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.720782] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf88dea-daae-4d5b-837c-d00e9edc482f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.759688] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a34f0d-61fc-4bc6-a33e-ef6022bd5d27 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.767626] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d5aeff-a113-494b-8246-8bd3c9d3d037 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.782034] env[61906]: DEBUG nova.compute.provider_tree [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.834982] env[61906]: DEBUG nova.compute.manager [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 790.878947] env[61906]: DEBUG nova.network.neutron [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Successfully created port: 47fd5cbf-baa2-424e-a003-1d4116844aad {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 790.883914] env[61906]: INFO nova.scheduler.client.report [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Deleted allocations for instance 45621a25-cd9c-4931-899c-647da0b50ec7 [ 791.080377] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Releasing lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.080727] env[61906]: DEBUG nova.compute.manager [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Instance network_info: |[{"id": "f38d9ad1-e3be-4227-981d-b2b14c6cb912", "address": "fa:16:3e:24:7a:87", "network": {"id": "1844d418-05f6-4457-83a6-08833666041a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1376388956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d5318e962dd4c17af474c9d0786a414", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed8a78a1-87dc-488e-a092-afd1c2a2ddde", "external-id": "nsx-vlan-transportzone-21", "segmentation_id": 21, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf38d9ad1-e3", "ovs_interfaceid": "f38d9ad1-e3be-4227-981d-b2b14c6cb912", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 791.081196] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:7a:87', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ed8a78a1-87dc-488e-a092-afd1c2a2ddde', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f38d9ad1-e3be-4227-981d-b2b14c6cb912', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 791.088639] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Creating folder: Project (5d5318e962dd4c17af474c9d0786a414). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 791.088920] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-02fb2048-10e5-47f1-913c-9e0cca5812f7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.099419] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Created folder: Project (5d5318e962dd4c17af474c9d0786a414) in parent group-v288914. [ 791.099618] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Creating folder: Instances. Parent ref: group-v288941. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 791.099858] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ba985e9-818d-4883-b662-94dfe3f553c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.108551] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Created folder: Instances in parent group-v288941. [ 791.109117] env[61906]: DEBUG oslo.service.loopingcall [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.109117] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 791.109265] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d4680b46-6c71-4b3d-b911-50ae340fd158 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.128065] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 791.128065] env[61906]: value = "task-1356527" [ 791.128065] env[61906]: _type = "Task" [ 791.128065] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.135908] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356527, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.161501] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356524, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.196958] env[61906]: DEBUG nova.network.neutron [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Successfully created port: 2bfbe252-2380-4a3c-a47d-39615dfa7991 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 791.290021] env[61906]: DEBUG nova.scheduler.client.report [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.350607] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Acquiring lock "0af2b07b-0632-4376-b502-58fe23d79f02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.350834] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Lock "0af2b07b-0632-4376-b502-58fe23d79f02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.394289] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cd0e6b9d-6be3-41f7-a415-5a3c304a4be6 tempest-ServerActionsV293TestJSON-1885527251 tempest-ServerActionsV293TestJSON-1885527251-project-member] Lock "45621a25-cd9c-4931-899c-647da0b50ec7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.467s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.639085] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356527, 'name': CreateVM_Task, 'duration_secs': 0.370822} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.639322] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 791.640072] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.640293] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.640640] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 791.640923] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cb0aec9-0a5e-45e4-9b5a-19148fcb5d3d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.645464] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for the task: (returnval){ [ 791.645464] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525eb2bb-87c0-f293-a8ed-c8081efb32c8" [ 791.645464] env[61906]: _type = "Task" [ 791.645464] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.654284] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525eb2bb-87c0-f293-a8ed-c8081efb32c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.662763] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356524, 'name': ReconfigVM_Task, 'duration_secs': 1.15835} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.664015] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 77791983-cef1-4f07-9b62-d52d335b08c8/77791983-cef1-4f07-9b62-d52d335b08c8.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.664659] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-73a680f3-c61e-4c95-b5cd-bba10d1d9c9b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.672269] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Waiting for the task: (returnval){ [ 791.672269] env[61906]: value = "task-1356528" [ 791.672269] env[61906]: _type = "Task" [ 791.672269] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.680866] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356528, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.798025] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.973s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.798025] env[61906]: ERROR nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 12ff7474-0485-4fe4-9e38-6cc050ac31f8, please check neutron logs for more information. [ 791.798025] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Traceback (most recent call last): [ 791.798025] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 791.798025] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] self.driver.spawn(context, instance, image_meta, [ 791.798025] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 791.798025] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 791.798025] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 791.798025] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] vm_ref = self.build_virtual_machine(instance, [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] vif_infos = vmwarevif.get_vif_info(self._session, [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] for vif in network_info: [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] return self._sync_wrapper(fn, *args, **kwargs) [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] self.wait() [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] self[:] = self._gt.wait() [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] return self._exit_event.wait() [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 791.798355] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] result = hub.switch() [ 791.798693] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 791.798693] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] return self.greenlet.switch() [ 791.798693] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.798693] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] result = function(*args, **kwargs) [ 791.798693] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.798693] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] return func(*args, **kwargs) [ 791.798693] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.798693] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] raise e [ 791.798693] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.798693] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] nwinfo = self.network_api.allocate_for_instance( [ 791.798693] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.798693] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] created_port_ids = self._update_ports_for_instance( [ 791.798693] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.798999] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] with excutils.save_and_reraise_exception(): [ 791.798999] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.798999] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] self.force_reraise() [ 791.798999] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.798999] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] raise self.value [ 791.798999] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.798999] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] updated_port = self._update_port( [ 791.798999] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.798999] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] _ensure_no_port_binding_failure(port) [ 791.798999] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.798999] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] raise exception.PortBindingFailed(port_id=port['id']) [ 791.798999] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] nova.exception.PortBindingFailed: Binding failed for port 12ff7474-0485-4fe4-9e38-6cc050ac31f8, please check neutron logs for more information. [ 791.798999] env[61906]: ERROR nova.compute.manager [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] [ 791.799367] env[61906]: DEBUG nova.compute.utils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Binding failed for port 12ff7474-0485-4fe4-9e38-6cc050ac31f8, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 791.800264] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.181s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.807242] env[61906]: DEBUG nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Build of instance db4eb67a-12dc-4347-ba29-1af46626a87d was re-scheduled: Binding failed for port 12ff7474-0485-4fe4-9e38-6cc050ac31f8, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 791.807593] env[61906]: DEBUG nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 791.807822] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Acquiring lock "refresh_cache-db4eb67a-12dc-4347-ba29-1af46626a87d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.807994] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Acquired lock "refresh_cache-db4eb67a-12dc-4347-ba29-1af46626a87d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.808330] env[61906]: DEBUG nova.network.neutron [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.847082] env[61906]: DEBUG nova.compute.manager [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 791.883129] env[61906]: DEBUG nova.virt.hardware [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 791.883129] env[61906]: DEBUG nova.virt.hardware [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 791.883129] env[61906]: DEBUG nova.virt.hardware [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 791.883129] env[61906]: DEBUG nova.virt.hardware [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 791.883309] env[61906]: DEBUG nova.virt.hardware [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 791.883309] env[61906]: DEBUG nova.virt.hardware [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 791.883309] env[61906]: DEBUG nova.virt.hardware [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 791.883309] env[61906]: DEBUG nova.virt.hardware [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 791.883309] env[61906]: DEBUG nova.virt.hardware [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 791.888045] env[61906]: DEBUG nova.virt.hardware [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 791.888333] env[61906]: DEBUG nova.virt.hardware [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 791.889503] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35dc1764-0bc7-463e-b495-d637467de111 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.897658] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac3c0cb-0f88-41d5-adbd-a8973e517ea0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.901416] env[61906]: DEBUG nova.compute.manager [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 791.938699] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquiring lock "a6e54432-a58e-49ba-a5c8-e6188f2ea4ff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.938960] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Lock "a6e54432-a58e-49ba-a5c8-e6188f2ea4ff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.939189] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquiring lock "a6e54432-a58e-49ba-a5c8-e6188f2ea4ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.939372] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Lock "a6e54432-a58e-49ba-a5c8-e6188f2ea4ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.939537] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Lock "a6e54432-a58e-49ba-a5c8-e6188f2ea4ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.941900] env[61906]: INFO nova.compute.manager [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Terminating instance [ 791.943604] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquiring lock "refresh_cache-a6e54432-a58e-49ba-a5c8-e6188f2ea4ff" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.943767] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquired lock "refresh_cache-a6e54432-a58e-49ba-a5c8-e6188f2ea4ff" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.943942] env[61906]: DEBUG nova.network.neutron [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 792.158399] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525eb2bb-87c0-f293-a8ed-c8081efb32c8, 'name': SearchDatastore_Task, 'duration_secs': 0.008912} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.158399] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.158635] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 792.158876] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.159049] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.159264] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 792.159527] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-59d3050f-4810-4d5b-ab36-97057394b956 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.168042] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 792.168251] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 792.168976] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f994868-7657-49ae-924f-482e252a38c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.177636] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for the task: (returnval){ [ 792.177636] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]524e1023-1d77-3c4e-e1ab-5f13fdfb6234" [ 792.177636] env[61906]: _type = "Task" [ 792.177636] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.184911] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356528, 'name': Rename_Task, 'duration_secs': 0.169204} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.185590] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 792.185882] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b048282-7fc0-4da9-863a-246d418b880f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.191173] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]524e1023-1d77-3c4e-e1ab-5f13fdfb6234, 'name': SearchDatastore_Task, 'duration_secs': 0.012606} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.192361] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cff364c4-20bd-4ae5-9e38-9b6f195d87cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.196157] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Waiting for the task: (returnval){ [ 792.196157] env[61906]: value = "task-1356529" [ 792.196157] env[61906]: _type = "Task" [ 792.196157] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.200360] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for the task: (returnval){ [ 792.200360] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529dbc9a-48cc-359c-1540-f6a73022edf8" [ 792.200360] env[61906]: _type = "Task" [ 792.200360] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.213433] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529dbc9a-48cc-359c-1540-f6a73022edf8, 'name': SearchDatastore_Task, 'duration_secs': 0.009701} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.216757] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.217145] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 7943ab1a-82df-48dd-874c-15dc015db51e/7943ab1a-82df-48dd-874c-15dc015db51e.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 792.217457] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356529, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.219307] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0beeaf25-10a1-4811-81bb-4f0255ee4248 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.224534] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for the task: (returnval){ [ 792.224534] env[61906]: value = "task-1356531" [ 792.224534] env[61906]: _type = "Task" [ 792.224534] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.234025] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356531, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.331157] env[61906]: DEBUG nova.network.neutron [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.433088] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.473193] env[61906]: DEBUG nova.network.neutron [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.505809] env[61906]: DEBUG nova.network.neutron [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.543017] env[61906]: DEBUG nova.network.neutron [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.652176] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89e8eaf-d85c-4a07-90e9-ab7685d44269 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.663025] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2d59ae-9326-4142-a0bb-375c965f09c7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.707978] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0de7700d-1574-46aa-8b39-2c670727342d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.722786] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0a7e1f-e5fc-4d8e-966e-b3d38df30531 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.727617] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356529, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.736875] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356531, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510401} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.745140] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 7943ab1a-82df-48dd-874c-15dc015db51e/7943ab1a-82df-48dd-874c-15dc015db51e.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 792.745140] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.745706] env[61906]: DEBUG nova.compute.provider_tree [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.747248] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0d9cf4c0-0864-476e-b185-93d08a871b56 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.754919] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for the task: (returnval){ [ 792.754919] env[61906]: value = "task-1356532" [ 792.754919] env[61906]: _type = "Task" [ 792.754919] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.762988] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356532, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.012021] env[61906]: DEBUG nova.compute.manager [req-6ac6df68-b95f-4c97-94d9-813ed889c28a req-fea6ea61-8d5f-4fa7-a815-e37819262718 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Received event network-vif-plugged-47fd5cbf-baa2-424e-a003-1d4116844aad {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.012268] env[61906]: DEBUG oslo_concurrency.lockutils [req-6ac6df68-b95f-4c97-94d9-813ed889c28a req-fea6ea61-8d5f-4fa7-a815-e37819262718 service nova] Acquiring lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.012842] env[61906]: DEBUG oslo_concurrency.lockutils [req-6ac6df68-b95f-4c97-94d9-813ed889c28a req-fea6ea61-8d5f-4fa7-a815-e37819262718 service nova] Lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.013108] env[61906]: DEBUG oslo_concurrency.lockutils [req-6ac6df68-b95f-4c97-94d9-813ed889c28a req-fea6ea61-8d5f-4fa7-a815-e37819262718 service nova] Lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.013326] env[61906]: DEBUG nova.compute.manager [req-6ac6df68-b95f-4c97-94d9-813ed889c28a req-fea6ea61-8d5f-4fa7-a815-e37819262718 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] No waiting events found dispatching network-vif-plugged-47fd5cbf-baa2-424e-a003-1d4116844aad {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 793.013530] env[61906]: WARNING nova.compute.manager [req-6ac6df68-b95f-4c97-94d9-813ed889c28a req-fea6ea61-8d5f-4fa7-a815-e37819262718 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Received unexpected event network-vif-plugged-47fd5cbf-baa2-424e-a003-1d4116844aad for instance with vm_state building and task_state spawning. [ 793.015304] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Releasing lock "refresh_cache-db4eb67a-12dc-4347-ba29-1af46626a87d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.015589] env[61906]: DEBUG nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 793.015765] env[61906]: DEBUG nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 793.015944] env[61906]: DEBUG nova.network.neutron [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 793.035906] env[61906]: DEBUG nova.network.neutron [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.046530] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Releasing lock "refresh_cache-a6e54432-a58e-49ba-a5c8-e6188f2ea4ff" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.046933] env[61906]: DEBUG nova.compute.manager [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 793.047207] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 793.048317] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deea231b-80de-48a5-9969-10fa433f7f61 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.057089] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 793.057359] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76f0ddec-92b5-421d-98e4-ae804fa80eab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.063985] env[61906]: DEBUG oslo_vmware.api [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 793.063985] env[61906]: value = "task-1356533" [ 793.063985] env[61906]: _type = "Task" [ 793.063985] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.069327] env[61906]: DEBUG nova.network.neutron [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Successfully updated port: 47fd5cbf-baa2-424e-a003-1d4116844aad {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 793.078487] env[61906]: DEBUG oslo_vmware.api [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356533, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.214120] env[61906]: DEBUG oslo_vmware.api [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356529, 'name': PowerOnVM_Task, 'duration_secs': 0.548352} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.214445] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 793.214670] env[61906]: INFO nova.compute.manager [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Took 8.25 seconds to spawn the instance on the hypervisor. [ 793.214908] env[61906]: DEBUG nova.compute.manager [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 793.215781] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5518d10-b771-49ff-989e-6e4e78cb10a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.252578] env[61906]: DEBUG nova.scheduler.client.report [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.263791] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356532, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069641} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.264184] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.265081] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e07d76-e31e-48be-8790-3c633cc40ba1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.288969] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 7943ab1a-82df-48dd-874c-15dc015db51e/7943ab1a-82df-48dd-874c-15dc015db51e.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.290168] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea58bdbd-7a35-4cc7-b212-67c8f9381a96 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.313311] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for the task: (returnval){ [ 793.313311] env[61906]: value = "task-1356534" [ 793.313311] env[61906]: _type = "Task" [ 793.313311] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.324506] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356534, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.537966] env[61906]: DEBUG nova.network.neutron [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.577205] env[61906]: DEBUG oslo_vmware.api [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356533, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.732284] env[61906]: INFO nova.compute.manager [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Took 27.83 seconds to build instance. [ 793.756898] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.956s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.757353] env[61906]: ERROR nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dd3c4c33-8c85-4958-9053-440584018c37, please check neutron logs for more information. [ 793.757353] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Traceback (most recent call last): [ 793.757353] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 793.757353] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] self.driver.spawn(context, instance, image_meta, [ 793.757353] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 793.757353] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] self._vmops.spawn(context, instance, image_meta, injected_files, [ 793.757353] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 793.757353] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] vm_ref = self.build_virtual_machine(instance, [ 793.757353] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 793.757353] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] vif_infos = vmwarevif.get_vif_info(self._session, [ 793.757353] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 793.757794] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] for vif in network_info: [ 793.757794] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 793.757794] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] return self._sync_wrapper(fn, *args, **kwargs) [ 793.757794] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 793.757794] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] self.wait() [ 793.757794] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 793.757794] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] self[:] = self._gt.wait() [ 793.757794] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 793.757794] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] return self._exit_event.wait() [ 793.757794] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 793.757794] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] result = hub.switch() [ 793.757794] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 793.757794] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] return self.greenlet.switch() [ 793.758228] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 793.758228] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] result = function(*args, **kwargs) [ 793.758228] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 793.758228] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] return func(*args, **kwargs) [ 793.758228] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 793.758228] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] raise e [ 793.758228] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.758228] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] nwinfo = self.network_api.allocate_for_instance( [ 793.758228] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.758228] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] created_port_ids = self._update_ports_for_instance( [ 793.758228] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.758228] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] with excutils.save_and_reraise_exception(): [ 793.758228] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.759987] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] self.force_reraise() [ 793.759987] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.759987] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] raise self.value [ 793.759987] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.759987] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] updated_port = self._update_port( [ 793.759987] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.759987] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] _ensure_no_port_binding_failure(port) [ 793.759987] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.759987] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] raise exception.PortBindingFailed(port_id=port['id']) [ 793.759987] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] nova.exception.PortBindingFailed: Binding failed for port dd3c4c33-8c85-4958-9053-440584018c37, please check neutron logs for more information. [ 793.759987] env[61906]: ERROR nova.compute.manager [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] [ 793.760379] env[61906]: DEBUG nova.compute.utils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Binding failed for port dd3c4c33-8c85-4958-9053-440584018c37, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 793.763243] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.365s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.764670] env[61906]: INFO nova.compute.claims [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.770573] env[61906]: DEBUG nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Build of instance b1429121-3c7e-4982-9059-5eb98a7f92df was re-scheduled: Binding failed for port dd3c4c33-8c85-4958-9053-440584018c37, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 793.772103] env[61906]: DEBUG nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 793.772376] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "refresh_cache-b1429121-3c7e-4982-9059-5eb98a7f92df" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.772774] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "refresh_cache-b1429121-3c7e-4982-9059-5eb98a7f92df" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.773452] env[61906]: DEBUG nova.network.neutron [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 793.826311] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356534, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.041294] env[61906]: INFO nova.compute.manager [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] [instance: db4eb67a-12dc-4347-ba29-1af46626a87d] Took 1.03 seconds to deallocate network for instance. [ 794.079627] env[61906]: DEBUG oslo_vmware.api [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356533, 'name': PowerOffVM_Task, 'duration_secs': 0.517426} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.079627] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 794.079627] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 794.079627] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7a842e95-e4fa-4db6-b08a-e0078e7228ac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.103913] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 794.104163] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 794.104339] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Deleting the datastore file [datastore1] a6e54432-a58e-49ba-a5c8-e6188f2ea4ff {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 794.104590] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63c56c5f-ebe5-419e-8cf1-e59e5cd45cdf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.112157] env[61906]: DEBUG oslo_vmware.api [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for the task: (returnval){ [ 794.112157] env[61906]: value = "task-1356536" [ 794.112157] env[61906]: _type = "Task" [ 794.112157] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.119838] env[61906]: DEBUG oslo_vmware.api [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356536, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.235586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3d1bf272-87f4-47d4-ba19-50934b865000 tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Lock "77791983-cef1-4f07-9b62-d52d335b08c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.598s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.304435] env[61906]: DEBUG nova.network.neutron [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.326779] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356534, 'name': ReconfigVM_Task, 'duration_secs': 0.958849} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.327085] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 7943ab1a-82df-48dd-874c-15dc015db51e/7943ab1a-82df-48dd-874c-15dc015db51e.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 794.327779] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87a7d475-d447-4f93-be38-17ffa68b3f95 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.336908] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for the task: (returnval){ [ 794.336908] env[61906]: value = "task-1356537" [ 794.336908] env[61906]: _type = "Task" [ 794.336908] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.350916] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356537, 'name': Rename_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.451865] env[61906]: DEBUG nova.network.neutron [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.622858] env[61906]: DEBUG oslo_vmware.api [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Task: {'id': task-1356536, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.107308} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.623120] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 794.623299] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 794.623474] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 794.623888] env[61906]: INFO nova.compute.manager [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Took 1.58 seconds to destroy the instance on the hypervisor. [ 794.624159] env[61906]: DEBUG oslo.service.loopingcall [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.624331] env[61906]: DEBUG nova.compute.manager [-] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.624419] env[61906]: DEBUG nova.network.neutron [-] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 794.649145] env[61906]: DEBUG nova.network.neutron [-] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.739008] env[61906]: DEBUG nova.compute.manager [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.807815] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Acquiring lock "77791983-cef1-4f07-9b62-d52d335b08c8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.808105] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Lock "77791983-cef1-4f07-9b62-d52d335b08c8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.808316] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Acquiring lock "77791983-cef1-4f07-9b62-d52d335b08c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.808500] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Lock "77791983-cef1-4f07-9b62-d52d335b08c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.808697] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Lock "77791983-cef1-4f07-9b62-d52d335b08c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.810894] env[61906]: INFO nova.compute.manager [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Terminating instance [ 794.815704] env[61906]: DEBUG nova.compute.manager [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 794.815704] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 794.816218] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87200a8-7215-4cba-85d3-2a33e4b3d756 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.823588] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 794.825983] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80ec7808-45bd-41e7-acf5-c175a4dd4979 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.833845] env[61906]: DEBUG oslo_vmware.api [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Waiting for the task: (returnval){ [ 794.833845] env[61906]: value = "task-1356538" [ 794.833845] env[61906]: _type = "Task" [ 794.833845] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.845925] env[61906]: DEBUG oslo_vmware.api [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356538, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.851388] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356537, 'name': Rename_Task, 'duration_secs': 0.232413} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.856107] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 794.856107] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-16462806-6f55-4dfc-8616-e952b80f1605 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.861555] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for the task: (returnval){ [ 794.861555] env[61906]: value = "task-1356539" [ 794.861555] env[61906]: _type = "Task" [ 794.861555] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.869327] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356539, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.956254] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "refresh_cache-b1429121-3c7e-4982-9059-5eb98a7f92df" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.956499] env[61906]: DEBUG nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 794.956689] env[61906]: DEBUG nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.957810] env[61906]: DEBUG nova.network.neutron [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 794.979638] env[61906]: DEBUG nova.network.neutron [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.035542] env[61906]: DEBUG nova.compute.manager [req-9814e0da-7b67-4f1f-ae73-1910ca409339 req-8d102292-cb20-468f-9c39-eda304262109 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Received event network-changed-47fd5cbf-baa2-424e-a003-1d4116844aad {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.035761] env[61906]: DEBUG nova.compute.manager [req-9814e0da-7b67-4f1f-ae73-1910ca409339 req-8d102292-cb20-468f-9c39-eda304262109 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Refreshing instance network info cache due to event network-changed-47fd5cbf-baa2-424e-a003-1d4116844aad. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 795.036159] env[61906]: DEBUG oslo_concurrency.lockutils [req-9814e0da-7b67-4f1f-ae73-1910ca409339 req-8d102292-cb20-468f-9c39-eda304262109 service nova] Acquiring lock "refresh_cache-2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.036263] env[61906]: DEBUG oslo_concurrency.lockutils [req-9814e0da-7b67-4f1f-ae73-1910ca409339 req-8d102292-cb20-468f-9c39-eda304262109 service nova] Acquired lock "refresh_cache-2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.036423] env[61906]: DEBUG nova.network.neutron [req-9814e0da-7b67-4f1f-ae73-1910ca409339 req-8d102292-cb20-468f-9c39-eda304262109 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Refreshing network info cache for port 47fd5cbf-baa2-424e-a003-1d4116844aad {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 795.072788] env[61906]: INFO nova.scheduler.client.report [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Deleted allocations for instance db4eb67a-12dc-4347-ba29-1af46626a87d [ 795.080794] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8c2f62-a4ce-4f1c-868c-5d2fd2b82177 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.091039] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1abb8bd3-eb01-44e1-a468-3f1219a67cab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.126771] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131a9539-147d-43ba-a612-4201785fe100 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.135543] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207798e5-f1ba-4ec3-9785-5b8538ace17f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.150444] env[61906]: DEBUG nova.compute.provider_tree [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.151725] env[61906]: DEBUG nova.network.neutron [-] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.277320] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.347841] env[61906]: DEBUG oslo_vmware.api [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356538, 'name': PowerOffVM_Task, 'duration_secs': 0.180043} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.348153] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 795.348325] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 795.348589] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf989656-f1d9-44e5-a5e3-a3dac4634ded {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.371542] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356539, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.414024] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 795.414247] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 795.414428] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Deleting the datastore file [datastore1] 77791983-cef1-4f07-9b62-d52d335b08c8 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 795.414707] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75e78f4c-a5ff-4725-98fd-93ec51c62fa9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.422631] env[61906]: DEBUG oslo_vmware.api [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Waiting for the task: (returnval){ [ 795.422631] env[61906]: value = "task-1356541" [ 795.422631] env[61906]: _type = "Task" [ 795.422631] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.431177] env[61906]: DEBUG oslo_vmware.api [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356541, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.484337] env[61906]: DEBUG nova.network.neutron [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.521912] env[61906]: DEBUG nova.network.neutron [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Successfully updated port: 2bfbe252-2380-4a3c-a47d-39615dfa7991 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 795.588901] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f46f8f08-0278-487d-ac5f-831cef3ab488 tempest-ServerTagsTestJSON-1336927781 tempest-ServerTagsTestJSON-1336927781-project-member] Lock "db4eb67a-12dc-4347-ba29-1af46626a87d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.683s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.656013] env[61906]: DEBUG nova.scheduler.client.report [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.659653] env[61906]: INFO nova.compute.manager [-] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Took 1.04 seconds to deallocate network for instance. [ 795.717199] env[61906]: DEBUG nova.network.neutron [req-9814e0da-7b67-4f1f-ae73-1910ca409339 req-8d102292-cb20-468f-9c39-eda304262109 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.804480] env[61906]: DEBUG nova.network.neutron [req-9814e0da-7b67-4f1f-ae73-1910ca409339 req-8d102292-cb20-468f-9c39-eda304262109 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.871936] env[61906]: DEBUG oslo_vmware.api [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356539, 'name': PowerOnVM_Task, 'duration_secs': 0.517183} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.872225] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 795.872418] env[61906]: INFO nova.compute.manager [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Took 8.35 seconds to spawn the instance on the hypervisor. [ 795.872599] env[61906]: DEBUG nova.compute.manager [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 795.873368] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d48c1872-bd17-413d-90cf-fb173a4413a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.933819] env[61906]: DEBUG oslo_vmware.api [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Task: {'id': task-1356541, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156425} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.934136] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 795.934335] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 795.934514] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 795.934684] env[61906]: INFO nova.compute.manager [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Took 1.12 seconds to destroy the instance on the hypervisor. [ 795.934974] env[61906]: DEBUG oslo.service.loopingcall [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.935199] env[61906]: DEBUG nova.compute.manager [-] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 795.935293] env[61906]: DEBUG nova.network.neutron [-] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 795.986393] env[61906]: INFO nova.compute.manager [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: b1429121-3c7e-4982-9059-5eb98a7f92df] Took 1.03 seconds to deallocate network for instance. [ 796.032896] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquiring lock "refresh_cache-2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.091942] env[61906]: DEBUG nova.compute.manager [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 796.162114] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.162650] env[61906]: DEBUG nova.compute.manager [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 796.165217] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.136s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.166637] env[61906]: INFO nova.compute.claims [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 796.169887] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.307251] env[61906]: DEBUG oslo_concurrency.lockutils [req-9814e0da-7b67-4f1f-ae73-1910ca409339 req-8d102292-cb20-468f-9c39-eda304262109 service nova] Releasing lock "refresh_cache-2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.307705] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquired lock "refresh_cache-2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.307945] env[61906]: DEBUG nova.network.neutron [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 796.392731] env[61906]: INFO nova.compute.manager [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Took 26.32 seconds to build instance. [ 796.614038] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.626689] env[61906]: DEBUG nova.network.neutron [-] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.674846] env[61906]: DEBUG nova.compute.utils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.676743] env[61906]: DEBUG nova.compute.manager [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 796.676743] env[61906]: DEBUG nova.network.neutron [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 796.733224] env[61906]: DEBUG nova.policy [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4a7d6ca72b14c9f8d6dc4750fc701be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1551c412835748b1ade7631d725af68a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 796.874450] env[61906]: DEBUG nova.network.neutron [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.895427] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d3a6dd4-89f0-449a-ba4a-e45487eaa832 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "7943ab1a-82df-48dd-874c-15dc015db51e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.670s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.026580] env[61906]: INFO nova.scheduler.client.report [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Deleted allocations for instance b1429121-3c7e-4982-9059-5eb98a7f92df [ 797.073549] env[61906]: DEBUG nova.compute.manager [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Received event network-vif-plugged-2bfbe252-2380-4a3c-a47d-39615dfa7991 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 797.073776] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] Acquiring lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.073987] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] Lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.075150] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] Lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.077598] env[61906]: DEBUG nova.compute.manager [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] No waiting events found dispatching network-vif-plugged-2bfbe252-2380-4a3c-a47d-39615dfa7991 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 797.077598] env[61906]: WARNING nova.compute.manager [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Received unexpected event network-vif-plugged-2bfbe252-2380-4a3c-a47d-39615dfa7991 for instance with vm_state building and task_state spawning. [ 797.077598] env[61906]: DEBUG nova.compute.manager [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Received event network-changed-2bfbe252-2380-4a3c-a47d-39615dfa7991 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 797.077598] env[61906]: DEBUG nova.compute.manager [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Refreshing instance network info cache due to event network-changed-2bfbe252-2380-4a3c-a47d-39615dfa7991. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 797.077598] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] Acquiring lock "refresh_cache-2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.131315] env[61906]: INFO nova.compute.manager [-] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Took 1.20 seconds to deallocate network for instance. [ 797.180427] env[61906]: DEBUG nova.compute.manager [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 797.212877] env[61906]: DEBUG nova.network.neutron [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Successfully created port: 56efd84b-d24d-49e4-870b-7361a46db687 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.403051] env[61906]: DEBUG nova.compute.manager [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 797.468767] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec375112-4a8d-4991-a2a7-4a2a8a50b777 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.479355] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e21318d-b2cc-49d0-badf-8599be4fe6c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.517797] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356bb676-da77-449f-8142-d6df1252ec94 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.526583] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d840fd0-e792-4fa9-9832-715ecb970da8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.541367] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f21fd493-a837-4cce-8e05-eefb08737c48 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "b1429121-3c7e-4982-9059-5eb98a7f92df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.142s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.541891] env[61906]: DEBUG nova.compute.provider_tree [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.615814] env[61906]: DEBUG nova.network.neutron [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Updating instance_info_cache with network_info: [{"id": "47fd5cbf-baa2-424e-a003-1d4116844aad", "address": "fa:16:3e:aa:43:93", "network": {"id": "d7b4d993-4571-4719-8ad5-ca4f8ccce968", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-139289865", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "473163e61f064b969326e695af5c54b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d2e4070-a78e-4d08-a104-b6312ab65577", "external-id": "nsx-vlan-transportzone-292", "segmentation_id": 292, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47fd5cbf-ba", "ovs_interfaceid": "47fd5cbf-baa2-424e-a003-1d4116844aad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2bfbe252-2380-4a3c-a47d-39615dfa7991", "address": "fa:16:3e:b9:c6:96", "network": {"id": "f17f1373-3280-425d-8f09-e152908c4f2f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-57515248", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.172", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "473163e61f064b969326e695af5c54b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92e4d027-e755-417b-8eea-9a8f24b85140", "external-id": "nsx-vlan-transportzone-756", "segmentation_id": 756, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bfbe252-23", "ovs_interfaceid": "2bfbe252-2380-4a3c-a47d-39615dfa7991", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.639342] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.933779] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.047905] env[61906]: DEBUG nova.scheduler.client.report [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 798.050479] env[61906]: DEBUG nova.compute.manager [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 798.123616] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Releasing lock "refresh_cache-2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.124219] env[61906]: DEBUG nova.compute.manager [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Instance network_info: |[{"id": "47fd5cbf-baa2-424e-a003-1d4116844aad", "address": "fa:16:3e:aa:43:93", "network": {"id": "d7b4d993-4571-4719-8ad5-ca4f8ccce968", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-139289865", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "473163e61f064b969326e695af5c54b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d2e4070-a78e-4d08-a104-b6312ab65577", "external-id": "nsx-vlan-transportzone-292", "segmentation_id": 292, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47fd5cbf-ba", "ovs_interfaceid": "47fd5cbf-baa2-424e-a003-1d4116844aad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2bfbe252-2380-4a3c-a47d-39615dfa7991", "address": "fa:16:3e:b9:c6:96", "network": {"id": "f17f1373-3280-425d-8f09-e152908c4f2f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-57515248", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.172", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "473163e61f064b969326e695af5c54b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92e4d027-e755-417b-8eea-9a8f24b85140", "external-id": "nsx-vlan-transportzone-756", "segmentation_id": 756, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bfbe252-23", "ovs_interfaceid": "2bfbe252-2380-4a3c-a47d-39615dfa7991", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 798.125534] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] Acquired lock "refresh_cache-2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.125534] env[61906]: DEBUG nova.network.neutron [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Refreshing network info cache for port 2bfbe252-2380-4a3c-a47d-39615dfa7991 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 798.127371] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:43:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d2e4070-a78e-4d08-a104-b6312ab65577', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '47fd5cbf-baa2-424e-a003-1d4116844aad', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:c6:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92e4d027-e755-417b-8eea-9a8f24b85140', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2bfbe252-2380-4a3c-a47d-39615dfa7991', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 798.138639] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Creating folder: Project (473163e61f064b969326e695af5c54b1). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 798.143164] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9547de0a-c42d-4790-b446-99f54aecfebb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.157577] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Created folder: Project (473163e61f064b969326e695af5c54b1) in parent group-v288914. [ 798.158082] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Creating folder: Instances. Parent ref: group-v288944. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 798.158382] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b74d25b4-9449-4405-acd3-8a165e895432 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.169443] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Created folder: Instances in parent group-v288944. [ 798.169771] env[61906]: DEBUG oslo.service.loopingcall [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.170015] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 798.170275] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b197e489-9744-4be0-9bfd-d4964d45e892 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.190708] env[61906]: DEBUG nova.compute.manager [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 798.195090] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 798.195090] env[61906]: value = "task-1356544" [ 798.195090] env[61906]: _type = "Task" [ 798.195090] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.208626] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356544, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.225424] env[61906]: DEBUG nova.virt.hardware [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.225810] env[61906]: DEBUG nova.virt.hardware [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.226118] env[61906]: DEBUG nova.virt.hardware [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.226445] env[61906]: DEBUG nova.virt.hardware [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.226720] env[61906]: DEBUG nova.virt.hardware [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.227185] env[61906]: DEBUG nova.virt.hardware [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.227379] env[61906]: DEBUG nova.virt.hardware [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.227671] env[61906]: DEBUG nova.virt.hardware [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.227980] env[61906]: DEBUG nova.virt.hardware [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.228316] env[61906]: DEBUG nova.virt.hardware [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.228622] env[61906]: DEBUG nova.virt.hardware [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.230069] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaeec29a-94c2-4209-8e10-1c02ce84e188 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.242453] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d59bc526-9c67-46e0-baef-7146058e1faa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.435897] env[61906]: DEBUG nova.network.neutron [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Updated VIF entry in instance network info cache for port 2bfbe252-2380-4a3c-a47d-39615dfa7991. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 798.436347] env[61906]: DEBUG nova.network.neutron [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Updating instance_info_cache with network_info: [{"id": "47fd5cbf-baa2-424e-a003-1d4116844aad", "address": "fa:16:3e:aa:43:93", "network": {"id": "d7b4d993-4571-4719-8ad5-ca4f8ccce968", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-139289865", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "473163e61f064b969326e695af5c54b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d2e4070-a78e-4d08-a104-b6312ab65577", "external-id": "nsx-vlan-transportzone-292", "segmentation_id": 292, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47fd5cbf-ba", "ovs_interfaceid": "47fd5cbf-baa2-424e-a003-1d4116844aad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2bfbe252-2380-4a3c-a47d-39615dfa7991", "address": "fa:16:3e:b9:c6:96", "network": {"id": "f17f1373-3280-425d-8f09-e152908c4f2f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-57515248", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.172", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "473163e61f064b969326e695af5c54b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92e4d027-e755-417b-8eea-9a8f24b85140", "external-id": "nsx-vlan-transportzone-756", "segmentation_id": 756, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bfbe252-23", "ovs_interfaceid": "2bfbe252-2380-4a3c-a47d-39615dfa7991", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.555296] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.390s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.555674] env[61906]: DEBUG nova.compute.manager [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 798.560418] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.577s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.562067] env[61906]: INFO nova.compute.claims [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.583260] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.610963] env[61906]: DEBUG oslo_concurrency.lockutils [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "7943ab1a-82df-48dd-874c-15dc015db51e" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.611231] env[61906]: DEBUG oslo_concurrency.lockutils [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "7943ab1a-82df-48dd-874c-15dc015db51e" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.611475] env[61906]: INFO nova.compute.manager [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Rebooting instance [ 798.709208] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356544, 'name': CreateVM_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.938685] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] Releasing lock "refresh_cache-2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.938946] env[61906]: DEBUG nova.compute.manager [req-5b41f2f6-8214-4c03-b79e-6f035bc63dc1 req-21fb5bbf-a502-4cbc-a263-4da63f06a121 service nova] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Received event network-vif-deleted-69416cc5-266f-493b-bc21-5288dbcec92e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 798.945690] env[61906]: DEBUG nova.network.neutron [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Successfully updated port: 56efd84b-d24d-49e4-870b-7361a46db687 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 799.062485] env[61906]: DEBUG nova.compute.utils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 799.063971] env[61906]: DEBUG nova.compute.manager [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 799.064160] env[61906]: DEBUG nova.network.neutron [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 799.101185] env[61906]: DEBUG nova.compute.manager [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Received event network-changed-f38d9ad1-e3be-4227-981d-b2b14c6cb912 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.101396] env[61906]: DEBUG nova.compute.manager [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Refreshing instance network info cache due to event network-changed-f38d9ad1-e3be-4227-981d-b2b14c6cb912. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 799.101609] env[61906]: DEBUG oslo_concurrency.lockutils [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] Acquiring lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.101751] env[61906]: DEBUG oslo_concurrency.lockutils [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] Acquired lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.101908] env[61906]: DEBUG nova.network.neutron [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Refreshing network info cache for port f38d9ad1-e3be-4227-981d-b2b14c6cb912 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 799.118137] env[61906]: DEBUG nova.policy [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf4fea47b929461785f4a487a56977a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3093466de87a45a2ac673c25bea19bc3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 799.135248] env[61906]: DEBUG oslo_concurrency.lockutils [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.211642] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356544, 'name': CreateVM_Task, 'duration_secs': 0.68945} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.211642] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 799.211642] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.211642] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.211642] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 799.211642] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29ce1c1c-fa08-4157-8620-b6426d2b3f49 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.215815] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Waiting for the task: (returnval){ [ 799.215815] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52594427-c866-f201-951b-faca8fde2649" [ 799.215815] env[61906]: _type = "Task" [ 799.215815] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.225185] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52594427-c866-f201-951b-faca8fde2649, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.319340] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "943c57f9-74c6-4465-9773-3fb01e78127b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.319570] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "943c57f9-74c6-4465-9773-3fb01e78127b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.394168] env[61906]: DEBUG nova.network.neutron [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Successfully created port: d962bee4-e780-4b00-b521-255e4c0e7fad {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 799.448637] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Acquiring lock "refresh_cache-8ab6b473-d13c-4341-9789-992ac3aba6a2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.448816] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Acquired lock "refresh_cache-8ab6b473-d13c-4341-9789-992ac3aba6a2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.448935] env[61906]: DEBUG nova.network.neutron [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 799.567444] env[61906]: DEBUG nova.compute.manager [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 799.729738] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52594427-c866-f201-951b-faca8fde2649, 'name': SearchDatastore_Task, 'duration_secs': 0.0091} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.730822] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.730822] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.730822] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.730822] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.730822] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.731292] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-66395249-0280-4381-8668-a979afb0858e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.739498] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.739678] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 799.740387] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36c6175f-80b4-437e-ae16-29fd3cf27626 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.746331] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Waiting for the task: (returnval){ [ 799.746331] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a7d234-bba1-50bd-3fcd-889ffd4f6ec4" [ 799.746331] env[61906]: _type = "Task" [ 799.746331] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.758575] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a7d234-bba1-50bd-3fcd-889ffd4f6ec4, 'name': SearchDatastore_Task, 'duration_secs': 0.007764} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.759309] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d924abe4-6514-4e72-83f2-8228ae0327e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.768019] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Waiting for the task: (returnval){ [ 799.768019] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52df5ef6-88eb-f75d-e51a-3fe875e635c2" [ 799.768019] env[61906]: _type = "Task" [ 799.768019] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.773894] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52df5ef6-88eb-f75d-e51a-3fe875e635c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.843426] env[61906]: DEBUG nova.network.neutron [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Updated VIF entry in instance network info cache for port f38d9ad1-e3be-4227-981d-b2b14c6cb912. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 799.848020] env[61906]: DEBUG nova.network.neutron [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Updating instance_info_cache with network_info: [{"id": "f38d9ad1-e3be-4227-981d-b2b14c6cb912", "address": "fa:16:3e:24:7a:87", "network": {"id": "1844d418-05f6-4457-83a6-08833666041a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1376388956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d5318e962dd4c17af474c9d0786a414", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed8a78a1-87dc-488e-a092-afd1c2a2ddde", "external-id": "nsx-vlan-transportzone-21", "segmentation_id": 21, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf38d9ad1-e3", "ovs_interfaceid": "f38d9ad1-e3be-4227-981d-b2b14c6cb912", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.858908] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89991107-17f3-4ae3-8fda-5825c0a8aa2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.868671] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77880c71-9bd7-4fbb-b69e-bc6de079c3af {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.902604] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e006362-0ffa-4f8b-93ea-c3dfe825e7e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.910522] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94924144-d450-4311-885d-402d7e834a96 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.926018] env[61906]: DEBUG nova.compute.provider_tree [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.991642] env[61906]: DEBUG nova.network.neutron [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.162828] env[61906]: DEBUG nova.network.neutron [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Updating instance_info_cache with network_info: [{"id": "56efd84b-d24d-49e4-870b-7361a46db687", "address": "fa:16:3e:86:7f:2e", "network": {"id": "259879ab-2358-4d5d-849d-94b16ea3e2b9", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-60659934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1551c412835748b1ade7631d725af68a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56efd84b-d2", "ovs_interfaceid": "56efd84b-d24d-49e4-870b-7361a46db687", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.276976] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52df5ef6-88eb-f75d-e51a-3fe875e635c2, 'name': SearchDatastore_Task, 'duration_secs': 0.009179} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.277340] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.277601] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc/2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 800.277856] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3bce4da3-40d8-40bd-b192-7690afdfa996 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.284463] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Waiting for the task: (returnval){ [ 800.284463] env[61906]: value = "task-1356545" [ 800.284463] env[61906]: _type = "Task" [ 800.284463] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.292318] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356545, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.346633] env[61906]: DEBUG oslo_concurrency.lockutils [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] Releasing lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.346886] env[61906]: DEBUG nova.compute.manager [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Received event network-vif-plugged-56efd84b-d24d-49e4-870b-7361a46db687 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.347109] env[61906]: DEBUG oslo_concurrency.lockutils [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] Acquiring lock "8ab6b473-d13c-4341-9789-992ac3aba6a2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.347328] env[61906]: DEBUG oslo_concurrency.lockutils [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] Lock "8ab6b473-d13c-4341-9789-992ac3aba6a2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.347493] env[61906]: DEBUG oslo_concurrency.lockutils [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] Lock "8ab6b473-d13c-4341-9789-992ac3aba6a2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.347659] env[61906]: DEBUG nova.compute.manager [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] No waiting events found dispatching network-vif-plugged-56efd84b-d24d-49e4-870b-7361a46db687 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 800.347827] env[61906]: WARNING nova.compute.manager [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Received unexpected event network-vif-plugged-56efd84b-d24d-49e4-870b-7361a46db687 for instance with vm_state building and task_state spawning. [ 800.347988] env[61906]: DEBUG nova.compute.manager [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Received event network-changed-56efd84b-d24d-49e4-870b-7361a46db687 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.348167] env[61906]: DEBUG nova.compute.manager [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Refreshing instance network info cache due to event network-changed-56efd84b-d24d-49e4-870b-7361a46db687. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 800.348332] env[61906]: DEBUG oslo_concurrency.lockutils [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] Acquiring lock "refresh_cache-8ab6b473-d13c-4341-9789-992ac3aba6a2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.348550] env[61906]: DEBUG oslo_concurrency.lockutils [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquired lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.348706] env[61906]: DEBUG nova.network.neutron [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 800.429861] env[61906]: DEBUG nova.scheduler.client.report [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.582697] env[61906]: DEBUG nova.compute.manager [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 800.615832] env[61906]: DEBUG nova.virt.hardware [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 800.616018] env[61906]: DEBUG nova.virt.hardware [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 800.616163] env[61906]: DEBUG nova.virt.hardware [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 800.616354] env[61906]: DEBUG nova.virt.hardware [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 800.616501] env[61906]: DEBUG nova.virt.hardware [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 800.616646] env[61906]: DEBUG nova.virt.hardware [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 800.616859] env[61906]: DEBUG nova.virt.hardware [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 800.617027] env[61906]: DEBUG nova.virt.hardware [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 800.617203] env[61906]: DEBUG nova.virt.hardware [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 800.617381] env[61906]: DEBUG nova.virt.hardware [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 800.617662] env[61906]: DEBUG nova.virt.hardware [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 800.618563] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b595ab9c-2be1-4fb1-9623-818d8d90b839 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.627404] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07eddf68-7a26-4fe0-be4b-7184464a1e24 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.665634] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Releasing lock "refresh_cache-8ab6b473-d13c-4341-9789-992ac3aba6a2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.666011] env[61906]: DEBUG nova.compute.manager [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Instance network_info: |[{"id": "56efd84b-d24d-49e4-870b-7361a46db687", "address": "fa:16:3e:86:7f:2e", "network": {"id": "259879ab-2358-4d5d-849d-94b16ea3e2b9", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-60659934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1551c412835748b1ade7631d725af68a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56efd84b-d2", "ovs_interfaceid": "56efd84b-d24d-49e4-870b-7361a46db687", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 800.666378] env[61906]: DEBUG oslo_concurrency.lockutils [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] Acquired lock "refresh_cache-8ab6b473-d13c-4341-9789-992ac3aba6a2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.666579] env[61906]: DEBUG nova.network.neutron [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Refreshing network info cache for port 56efd84b-d24d-49e4-870b-7361a46db687 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 800.667865] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:7f:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e6f11c0d-c73a-47f5-b02e-47bff48da0e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '56efd84b-d24d-49e4-870b-7361a46db687', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.676630] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Creating folder: Project (1551c412835748b1ade7631d725af68a). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 800.677469] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ab419cee-d262-4e0c-a610-24c395614404 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.689642] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Created folder: Project (1551c412835748b1ade7631d725af68a) in parent group-v288914. [ 800.689801] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Creating folder: Instances. Parent ref: group-v288947. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 800.690045] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7bae3119-8164-4b76-ac3c-cc3111bf01c2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.703216] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Created folder: Instances in parent group-v288947. [ 800.703483] env[61906]: DEBUG oslo.service.loopingcall [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.703673] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 800.703889] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f58471a-0af5-497f-950a-b53cfb41047e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.723987] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.723987] env[61906]: value = "task-1356548" [ 800.723987] env[61906]: _type = "Task" [ 800.723987] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.732545] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356548, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.794986] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356545, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49149} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.795348] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc/2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 800.795618] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 800.795985] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7fe2b2a9-0b7b-496d-b7c9-fa3128e75a8c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.803378] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Waiting for the task: (returnval){ [ 800.803378] env[61906]: value = "task-1356549" [ 800.803378] env[61906]: _type = "Task" [ 800.803378] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.812744] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356549, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.936652] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.937386] env[61906]: DEBUG nova.compute.manager [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.941442] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.208s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.078119] env[61906]: DEBUG nova.network.neutron [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Successfully updated port: d962bee4-e780-4b00-b521-255e4c0e7fad {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 801.157064] env[61906]: DEBUG nova.network.neutron [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Updating instance_info_cache with network_info: [{"id": "f38d9ad1-e3be-4227-981d-b2b14c6cb912", "address": "fa:16:3e:24:7a:87", "network": {"id": "1844d418-05f6-4457-83a6-08833666041a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1376388956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d5318e962dd4c17af474c9d0786a414", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ed8a78a1-87dc-488e-a092-afd1c2a2ddde", "external-id": "nsx-vlan-transportzone-21", "segmentation_id": 21, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf38d9ad1-e3", "ovs_interfaceid": "f38d9ad1-e3be-4227-981d-b2b14c6cb912", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.230732] env[61906]: DEBUG nova.compute.manager [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Received event network-vif-plugged-d962bee4-e780-4b00-b521-255e4c0e7fad {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.230982] env[61906]: DEBUG oslo_concurrency.lockutils [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] Acquiring lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.231218] env[61906]: DEBUG oslo_concurrency.lockutils [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] Lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.231427] env[61906]: DEBUG oslo_concurrency.lockutils [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] Lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.231611] env[61906]: DEBUG nova.compute.manager [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] No waiting events found dispatching network-vif-plugged-d962bee4-e780-4b00-b521-255e4c0e7fad {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 801.231780] env[61906]: WARNING nova.compute.manager [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Received unexpected event network-vif-plugged-d962bee4-e780-4b00-b521-255e4c0e7fad for instance with vm_state building and task_state spawning. [ 801.231934] env[61906]: DEBUG nova.compute.manager [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Received event network-changed-d962bee4-e780-4b00-b521-255e4c0e7fad {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.232099] env[61906]: DEBUG nova.compute.manager [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Refreshing instance network info cache due to event network-changed-d962bee4-e780-4b00-b521-255e4c0e7fad. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 801.232288] env[61906]: DEBUG oslo_concurrency.lockutils [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] Acquiring lock "refresh_cache-9a23071d-40fd-4446-aa03-ecbddcafe3d6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.232424] env[61906]: DEBUG oslo_concurrency.lockutils [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] Acquired lock "refresh_cache-9a23071d-40fd-4446-aa03-ecbddcafe3d6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.232579] env[61906]: DEBUG nova.network.neutron [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Refreshing network info cache for port d962bee4-e780-4b00-b521-255e4c0e7fad {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 801.240690] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356548, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.314542] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356549, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.119506} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.315019] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.316190] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373cce85-4bb8-45c8-bd56-b868d3d94833 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.342499] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc/2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.345832] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0ce4663-8422-4c1d-936b-43c3112876c5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.366879] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Waiting for the task: (returnval){ [ 801.366879] env[61906]: value = "task-1356550" [ 801.366879] env[61906]: _type = "Task" [ 801.366879] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.376229] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356550, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.443144] env[61906]: DEBUG nova.network.neutron [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Updated VIF entry in instance network info cache for port 56efd84b-d24d-49e4-870b-7361a46db687. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 801.443532] env[61906]: DEBUG nova.network.neutron [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Updating instance_info_cache with network_info: [{"id": "56efd84b-d24d-49e4-870b-7361a46db687", "address": "fa:16:3e:86:7f:2e", "network": {"id": "259879ab-2358-4d5d-849d-94b16ea3e2b9", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-60659934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1551c412835748b1ade7631d725af68a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56efd84b-d2", "ovs_interfaceid": "56efd84b-d24d-49e4-870b-7361a46db687", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.450750] env[61906]: DEBUG nova.compute.utils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 801.452429] env[61906]: DEBUG nova.compute.manager [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 801.452429] env[61906]: DEBUG nova.network.neutron [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 801.495018] env[61906]: DEBUG nova.policy [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73219f5e96f64d1da86a19da9b18b06a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5533be74874b4094b7ef0f98a121f60b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.586677] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "refresh_cache-9a23071d-40fd-4446-aa03-ecbddcafe3d6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.663192] env[61906]: DEBUG oslo_concurrency.lockutils [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Releasing lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.665383] env[61906]: DEBUG nova.compute.manager [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 801.666515] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59deee19-289c-458f-a7fb-c5546478ab4b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.729871] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbed6eeb-537d-45b5-9142-0adf58f51a0d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.738311] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356548, 'name': CreateVM_Task, 'duration_secs': 0.616916} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.740178] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 801.740915] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.741113] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.741483] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 801.742380] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec266167-bf96-48ea-b4a7-29c9474f8bbc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.747766] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f662fc6-0f32-4a31-bfb1-ed629f6ac2a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.752954] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Waiting for the task: (returnval){ [ 801.752954] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521461cc-477f-74f4-3155-e45dd9478dd1" [ 801.752954] env[61906]: _type = "Task" [ 801.752954] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.784110] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513cc33f-9a94-4c09-9666-933c7983a109 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.799610] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521461cc-477f-74f4-3155-e45dd9478dd1, 'name': SearchDatastore_Task, 'duration_secs': 0.018058} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.799610] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.799610] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.799610] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.799610] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.799610] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.799610] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ed94ecf-0b27-46c5-a151-ce55a42c493f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.800722] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc1beec-86b1-432e-a7bd-1b802941f1d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.805886] env[61906]: DEBUG nova.network.neutron [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.809913] env[61906]: DEBUG nova.network.neutron [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Successfully created port: 76876157-42ef-4519-88b2-ca9a600e3b41 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.821248] env[61906]: DEBUG nova.compute.provider_tree [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 801.826072] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.826072] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 801.826072] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-710974ab-2bc0-492c-913e-3d78bf69d7bc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.832529] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Waiting for the task: (returnval){ [ 801.832529] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522afabf-b8c9-8179-0d69-c832ecf75288" [ 801.832529] env[61906]: _type = "Task" [ 801.832529] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.841845] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522afabf-b8c9-8179-0d69-c832ecf75288, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.880873] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356550, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.895271] env[61906]: DEBUG nova.network.neutron [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.946854] env[61906]: DEBUG oslo_concurrency.lockutils [req-a93b6c4b-805b-4c32-8ac9-758cc419019e req-cfc44a9f-99d3-47d0-9239-142b3c16731c service nova] Releasing lock "refresh_cache-8ab6b473-d13c-4341-9789-992ac3aba6a2" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.954628] env[61906]: DEBUG nova.compute.manager [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 802.345570] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522afabf-b8c9-8179-0d69-c832ecf75288, 'name': SearchDatastore_Task, 'duration_secs': 0.055791} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.345570] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b3f534b-ea2c-4c5d-a8b9-ef30214e5e38 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.350268] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Waiting for the task: (returnval){ [ 802.350268] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528b2e82-d67a-a0b8-1000-14b2c18f1ac3" [ 802.350268] env[61906]: _type = "Task" [ 802.350268] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.353907] env[61906]: ERROR nova.scheduler.client.report [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [req-f9116a2d-4b75-4f65-96c9-c65e8b0b878c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6e524f5d-985e-41eb-a7ae-996be3ae1c1a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f9116a2d-4b75-4f65-96c9-c65e8b0b878c"}]}: nova.exception.PortBindingFailed: Binding failed for port 52f119bb-d26b-495b-945b-9350e74ffb03, please check neutron logs for more information. [ 802.362356] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528b2e82-d67a-a0b8-1000-14b2c18f1ac3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.373090] env[61906]: DEBUG nova.scheduler.client.report [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Refreshing inventories for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 802.379362] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356550, 'name': ReconfigVM_Task, 'duration_secs': 0.993601} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.379920] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc/2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.380661] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-633e5343-f9c9-4ad7-9d2c-e8dbb33f9cfe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.387021] env[61906]: DEBUG nova.scheduler.client.report [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Updating ProviderTree inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 802.387021] env[61906]: DEBUG nova.compute.provider_tree [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 802.390290] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Waiting for the task: (returnval){ [ 802.390290] env[61906]: value = "task-1356551" [ 802.390290] env[61906]: _type = "Task" [ 802.390290] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.401695] env[61906]: DEBUG oslo_concurrency.lockutils [req-41c53d51-06eb-4dde-abe0-4df23307c4c4 req-b6b04478-747b-407b-988a-b6e5d924f249 service nova] Releasing lock "refresh_cache-9a23071d-40fd-4446-aa03-ecbddcafe3d6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.402072] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356551, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.402982] env[61906]: DEBUG nova.scheduler.client.report [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Refreshing aggregate associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 802.404985] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "refresh_cache-9a23071d-40fd-4446-aa03-ecbddcafe3d6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.405178] env[61906]: DEBUG nova.network.neutron [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 802.425056] env[61906]: DEBUG nova.scheduler.client.report [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Refreshing trait associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 802.652715] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1af7c64-1847-4a4a-9c32-b7525df74ac4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.660058] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd2223a-06aa-4848-a7bc-21a382e4d357 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.690175] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-787fd0a4-62cd-41c1-bf39-45f5c7d2a9dc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.693101] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51778210-a1e5-4283-bad5-7578d1ff89a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.702044] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed58b2f9-36ec-4621-927d-ad2dc18a2dcb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.705642] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Doing hard reboot of VM {{(pid=61906) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 802.706409] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-394cdea5-be52-40a3-99ff-22cc514b5d79 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.716825] env[61906]: DEBUG nova.compute.provider_tree [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 802.719892] env[61906]: DEBUG oslo_vmware.api [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for the task: (returnval){ [ 802.719892] env[61906]: value = "task-1356552" [ 802.719892] env[61906]: _type = "Task" [ 802.719892] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.727043] env[61906]: DEBUG oslo_vmware.api [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356552, 'name': ResetVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.861200] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528b2e82-d67a-a0b8-1000-14b2c18f1ac3, 'name': SearchDatastore_Task, 'duration_secs': 0.010943} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.861492] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.861756] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 8ab6b473-d13c-4341-9789-992ac3aba6a2/8ab6b473-d13c-4341-9789-992ac3aba6a2.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 802.862023] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c8a502bb-9d56-44d7-a24a-e97bbda242a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.868420] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Waiting for the task: (returnval){ [ 802.868420] env[61906]: value = "task-1356553" [ 802.868420] env[61906]: _type = "Task" [ 802.868420] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.876571] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356553, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.898593] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356551, 'name': Rename_Task, 'duration_secs': 0.157859} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.898853] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 802.899132] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2580b61e-13d2-4175-8f6b-2aaad82bd1ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.905552] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Waiting for the task: (returnval){ [ 802.905552] env[61906]: value = "task-1356554" [ 802.905552] env[61906]: _type = "Task" [ 802.905552] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.914091] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356554, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.941287] env[61906]: DEBUG nova.network.neutron [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.964682] env[61906]: DEBUG nova.compute.manager [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.990682] env[61906]: DEBUG nova.virt.hardware [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.990682] env[61906]: DEBUG nova.virt.hardware [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.990929] env[61906]: DEBUG nova.virt.hardware [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.991041] env[61906]: DEBUG nova.virt.hardware [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.991219] env[61906]: DEBUG nova.virt.hardware [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.991353] env[61906]: DEBUG nova.virt.hardware [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.991581] env[61906]: DEBUG nova.virt.hardware [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.991744] env[61906]: DEBUG nova.virt.hardware [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.991923] env[61906]: DEBUG nova.virt.hardware [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.992131] env[61906]: DEBUG nova.virt.hardware [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.992342] env[61906]: DEBUG nova.virt.hardware [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.993623] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4146dfd2-2be0-4ae0-9e9a-7c52f7d8b585 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.001528] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12cefe30-e499-42a4-83d9-70a27cc11e78 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.107332] env[61906]: DEBUG nova.network.neutron [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Updating instance_info_cache with network_info: [{"id": "d962bee4-e780-4b00-b521-255e4c0e7fad", "address": "fa:16:3e:76:3f:2b", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd962bee4-e7", "ovs_interfaceid": "d962bee4-e780-4b00-b521-255e4c0e7fad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.235908] env[61906]: DEBUG oslo_vmware.api [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356552, 'name': ResetVM_Task, 'duration_secs': 0.134343} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.235908] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Did hard reboot of VM {{(pid=61906) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 803.235908] env[61906]: DEBUG nova.compute.manager [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.235908] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c6a6d7-03b8-4715-91c6-263ed2fc5efe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.251463] env[61906]: DEBUG nova.scheduler.client.report [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Updated inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with generation 85 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 803.251621] env[61906]: DEBUG nova.compute.provider_tree [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Updating resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a generation from 85 to 86 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 803.251702] env[61906]: DEBUG nova.compute.provider_tree [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 803.378915] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356553, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489722} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.379186] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 8ab6b473-d13c-4341-9789-992ac3aba6a2/8ab6b473-d13c-4341-9789-992ac3aba6a2.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 803.379466] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.379731] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fda082e8-9e65-4f9a-b627-5c912a06cd1c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.385185] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Waiting for the task: (returnval){ [ 803.385185] env[61906]: value = "task-1356555" [ 803.385185] env[61906]: _type = "Task" [ 803.385185] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.393273] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356555, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.414009] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356554, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.428292] env[61906]: DEBUG nova.compute.manager [req-8942d72d-8d41-41ef-bb9d-9799b9a023ff req-dd9dc7fb-c58d-42b2-82c2-de8281d5b7c6 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Received event network-vif-plugged-76876157-42ef-4519-88b2-ca9a600e3b41 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 803.428505] env[61906]: DEBUG oslo_concurrency.lockutils [req-8942d72d-8d41-41ef-bb9d-9799b9a023ff req-dd9dc7fb-c58d-42b2-82c2-de8281d5b7c6 service nova] Acquiring lock "5adaa660-b736-4c11-9141-846cf475ccd5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.428715] env[61906]: DEBUG oslo_concurrency.lockutils [req-8942d72d-8d41-41ef-bb9d-9799b9a023ff req-dd9dc7fb-c58d-42b2-82c2-de8281d5b7c6 service nova] Lock "5adaa660-b736-4c11-9141-846cf475ccd5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.428887] env[61906]: DEBUG oslo_concurrency.lockutils [req-8942d72d-8d41-41ef-bb9d-9799b9a023ff req-dd9dc7fb-c58d-42b2-82c2-de8281d5b7c6 service nova] Lock "5adaa660-b736-4c11-9141-846cf475ccd5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.429145] env[61906]: DEBUG nova.compute.manager [req-8942d72d-8d41-41ef-bb9d-9799b9a023ff req-dd9dc7fb-c58d-42b2-82c2-de8281d5b7c6 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] No waiting events found dispatching network-vif-plugged-76876157-42ef-4519-88b2-ca9a600e3b41 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 803.429233] env[61906]: WARNING nova.compute.manager [req-8942d72d-8d41-41ef-bb9d-9799b9a023ff req-dd9dc7fb-c58d-42b2-82c2-de8281d5b7c6 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Received unexpected event network-vif-plugged-76876157-42ef-4519-88b2-ca9a600e3b41 for instance with vm_state building and task_state spawning. [ 803.553016] env[61906]: DEBUG nova.network.neutron [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Successfully updated port: 76876157-42ef-4519-88b2-ca9a600e3b41 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 803.609734] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "refresh_cache-9a23071d-40fd-4446-aa03-ecbddcafe3d6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.610077] env[61906]: DEBUG nova.compute.manager [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Instance network_info: |[{"id": "d962bee4-e780-4b00-b521-255e4c0e7fad", "address": "fa:16:3e:76:3f:2b", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd962bee4-e7", "ovs_interfaceid": "d962bee4-e780-4b00-b521-255e4c0e7fad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 803.610766] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:3f:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30f1dacf-8988-41b8-aa8f-e9530f65ef46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd962bee4-e780-4b00-b521-255e4c0e7fad', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 803.617949] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Creating folder: Project (3093466de87a45a2ac673c25bea19bc3). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 803.618256] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e938c815-bef1-44ca-b4dc-59e9f490acd7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.628090] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Created folder: Project (3093466de87a45a2ac673c25bea19bc3) in parent group-v288914. [ 803.628281] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Creating folder: Instances. Parent ref: group-v288950. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 803.628505] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9aed0e82-2ddb-445b-9c37-15ce96c7083f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.637326] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Created folder: Instances in parent group-v288950. [ 803.637402] env[61906]: DEBUG oslo.service.loopingcall [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 803.637596] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 803.637788] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-76f3d3f9-7034-4bc1-8e85-446375afc2bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.656743] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 803.656743] env[61906]: value = "task-1356558" [ 803.656743] env[61906]: _type = "Task" [ 803.656743] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.663985] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356558, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.751123] env[61906]: DEBUG oslo_concurrency.lockutils [None req-14104bde-bdce-4ac5-a2b5-7933280b20f7 tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "7943ab1a-82df-48dd-874c-15dc015db51e" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.140s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.756790] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.815s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.757797] env[61906]: ERROR nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 52f119bb-d26b-495b-945b-9350e74ffb03, please check neutron logs for more information. [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Traceback (most recent call last): [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] self.driver.spawn(context, instance, image_meta, [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] self._vmops.spawn(context, instance, image_meta, injected_files, [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] vm_ref = self.build_virtual_machine(instance, [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] vif_infos = vmwarevif.get_vif_info(self._session, [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] for vif in network_info: [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] return self._sync_wrapper(fn, *args, **kwargs) [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] self.wait() [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] self[:] = self._gt.wait() [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] return self._exit_event.wait() [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] current.throw(*self._exc) [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] result = function(*args, **kwargs) [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] return func(*args, **kwargs) [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] raise e [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] nwinfo = self.network_api.allocate_for_instance( [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] created_port_ids = self._update_ports_for_instance( [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] with excutils.save_and_reraise_exception(): [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] self.force_reraise() [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] raise self.value [ 803.757797] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 803.758995] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] updated_port = self._update_port( [ 803.758995] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 803.758995] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] _ensure_no_port_binding_failure(port) [ 803.758995] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 803.758995] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] raise exception.PortBindingFailed(port_id=port['id']) [ 803.758995] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] nova.exception.PortBindingFailed: Binding failed for port 52f119bb-d26b-495b-945b-9350e74ffb03, please check neutron logs for more information. [ 803.758995] env[61906]: ERROR nova.compute.manager [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] [ 803.758995] env[61906]: DEBUG nova.compute.utils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Binding failed for port 52f119bb-d26b-495b-945b-9350e74ffb03, please check neutron logs for more information. {{(pid=61906) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 803.759292] env[61906]: DEBUG oslo_concurrency.lockutils [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 13.070s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.759488] env[61906]: DEBUG nova.objects.instance [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61906) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 803.762985] env[61906]: DEBUG nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Build of instance c3cb1a2d-e905-497c-a4de-8c6579c37156 was re-scheduled: Binding failed for port 52f119bb-d26b-495b-945b-9350e74ffb03, please check neutron logs for more information. {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 803.762985] env[61906]: DEBUG nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Unplugging VIFs for instance {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 803.763671] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquiring lock "refresh_cache-c3cb1a2d-e905-497c-a4de-8c6579c37156" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.763671] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Acquired lock "refresh_cache-c3cb1a2d-e905-497c-a4de-8c6579c37156" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.763671] env[61906]: DEBUG nova.network.neutron [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.895202] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356555, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060639} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.895479] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 803.896254] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c399c5-8e33-49b4-86a6-6dd81ce9b366 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.919034] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 8ab6b473-d13c-4341-9789-992ac3aba6a2/8ab6b473-d13c-4341-9789-992ac3aba6a2.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.921972] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87e230e1-8768-42f8-bd62-68fa152d559a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.943347] env[61906]: DEBUG oslo_vmware.api [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356554, 'name': PowerOnVM_Task, 'duration_secs': 0.909134} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.944845] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 803.945111] env[61906]: INFO nova.compute.manager [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Took 12.10 seconds to spawn the instance on the hypervisor. [ 803.945328] env[61906]: DEBUG nova.compute.manager [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.946086] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Waiting for the task: (returnval){ [ 803.946086] env[61906]: value = "task-1356559" [ 803.946086] env[61906]: _type = "Task" [ 803.946086] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.946992] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5bf87e-dc5d-402d-82b6-ea3418b026e9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.959160] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356559, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.055671] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.055843] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.056075] env[61906]: DEBUG nova.network.neutron [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 804.167565] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356558, 'name': CreateVM_Task, 'duration_secs': 0.406765} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.167780] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 804.168518] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.169829] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.169829] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 804.169829] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b07c4b86-3866-4e31-9d4d-d989af337ebf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.176405] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 804.176405] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525f1fa8-e319-0396-09d7-f82cca950231" [ 804.176405] env[61906]: _type = "Task" [ 804.176405] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.182018] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525f1fa8-e319-0396-09d7-f82cca950231, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.291823] env[61906]: DEBUG nova.network.neutron [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.394502] env[61906]: DEBUG nova.network.neutron [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.460437] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356559, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.474150] env[61906]: INFO nova.compute.manager [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Took 32.51 seconds to build instance. [ 804.587545] env[61906]: DEBUG nova.network.neutron [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.690338] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525f1fa8-e319-0396-09d7-f82cca950231, 'name': SearchDatastore_Task, 'duration_secs': 0.012565} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.690658] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.690890] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 804.691182] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.691334] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.691519] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 804.691780] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7462ed6-d815-48a9-9723-ba56905d4ecb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.707179] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 804.707466] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 804.708146] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78212ef8-b6f3-4385-9a59-e22b4239dc09 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.713716] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 804.713716] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521f9a15-7fc9-cd57-d1cf-dc4502d2e394" [ 804.713716] env[61906]: _type = "Task" [ 804.713716] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.722226] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521f9a15-7fc9-cd57-d1cf-dc4502d2e394, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.774787] env[61906]: DEBUG oslo_concurrency.lockutils [None req-146ab7d2-d20f-4c39-8170-f22f370c9f66 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.776013] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.343s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.777584] env[61906]: INFO nova.compute.claims [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 804.807626] env[61906]: DEBUG nova.network.neutron [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Updating instance_info_cache with network_info: [{"id": "76876157-42ef-4519-88b2-ca9a600e3b41", "address": "fa:16:3e:66:6f:62", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76876157-42", "ovs_interfaceid": "76876157-42ef-4519-88b2-ca9a600e3b41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.872833] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "7943ab1a-82df-48dd-874c-15dc015db51e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.873159] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "7943ab1a-82df-48dd-874c-15dc015db51e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.873397] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "7943ab1a-82df-48dd-874c-15dc015db51e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.873590] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "7943ab1a-82df-48dd-874c-15dc015db51e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.873760] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "7943ab1a-82df-48dd-874c-15dc015db51e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.877562] env[61906]: INFO nova.compute.manager [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Terminating instance [ 804.879292] env[61906]: DEBUG nova.compute.manager [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 804.879482] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 804.880303] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab764d8b-b1df-4f93-929b-0b28c0cd8131 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.887796] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 804.888050] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3725f22e-b585-4fbb-a8d6-1426b3cf1f16 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.894036] env[61906]: DEBUG oslo_vmware.api [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for the task: (returnval){ [ 804.894036] env[61906]: value = "task-1356560" [ 804.894036] env[61906]: _type = "Task" [ 804.894036] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.897242] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Releasing lock "refresh_cache-c3cb1a2d-e905-497c-a4de-8c6579c37156" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.897453] env[61906]: DEBUG nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61906) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 804.897631] env[61906]: DEBUG nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 804.897802] env[61906]: DEBUG nova.network.neutron [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 804.903921] env[61906]: DEBUG oslo_vmware.api [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356560, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.913408] env[61906]: DEBUG nova.network.neutron [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.960852] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356559, 'name': ReconfigVM_Task, 'duration_secs': 0.533087} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.961144] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 8ab6b473-d13c-4341-9789-992ac3aba6a2/8ab6b473-d13c-4341-9789-992ac3aba6a2.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.961793] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-48e24ec3-db79-41fa-92c4-79c24c333f6d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.967992] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Waiting for the task: (returnval){ [ 804.967992] env[61906]: value = "task-1356561" [ 804.967992] env[61906]: _type = "Task" [ 804.967992] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.976215] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c56ef112-804f-444b-a4bb-a9e1f5e7cdf4 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.427s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.976455] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356561, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.224113] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521f9a15-7fc9-cd57-d1cf-dc4502d2e394, 'name': SearchDatastore_Task, 'duration_secs': 0.0091} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.224912] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a984e583-65b4-456a-9932-5578cc9ea29d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.230246] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 805.230246] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525be344-e1d4-87c6-9e1f-29a52b797b9e" [ 805.230246] env[61906]: _type = "Task" [ 805.230246] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.237802] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525be344-e1d4-87c6-9e1f-29a52b797b9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.309848] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.310183] env[61906]: DEBUG nova.compute.manager [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Instance network_info: |[{"id": "76876157-42ef-4519-88b2-ca9a600e3b41", "address": "fa:16:3e:66:6f:62", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76876157-42", "ovs_interfaceid": "76876157-42ef-4519-88b2-ca9a600e3b41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 805.310621] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:6f:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '76876157-42ef-4519-88b2-ca9a600e3b41', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.318133] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Creating folder: Project (5533be74874b4094b7ef0f98a121f60b). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.318399] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-19cb5f49-5686-490a-b2a2-4df067d135d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.327777] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquiring lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.328413] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.328413] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquiring lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.328555] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.328674] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.331934] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Created folder: Project (5533be74874b4094b7ef0f98a121f60b) in parent group-v288914. [ 805.332034] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Creating folder: Instances. Parent ref: group-v288953. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.332836] env[61906]: INFO nova.compute.manager [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Terminating instance [ 805.333919] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-07881e5b-cb61-4b1e-8b7b-c959dff4fbec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.335928] env[61906]: DEBUG nova.compute.manager [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 805.336310] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 805.337126] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb4056e-a869-436d-9474-2d92bfb3ba61 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.345936] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 805.346206] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c922d0f-661c-44b6-80b6-84048fbad7ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.348983] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Created folder: Instances in parent group-v288953. [ 805.349230] env[61906]: DEBUG oslo.service.loopingcall [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.349405] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 805.349588] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cedf2a2a-c7af-40ac-bdfd-6617811fdb37 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.365430] env[61906]: DEBUG oslo_vmware.api [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Waiting for the task: (returnval){ [ 805.365430] env[61906]: value = "task-1356564" [ 805.365430] env[61906]: _type = "Task" [ 805.365430] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.369651] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.369651] env[61906]: value = "task-1356565" [ 805.369651] env[61906]: _type = "Task" [ 805.369651] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.375591] env[61906]: DEBUG oslo_vmware.api [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356564, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.379932] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356565, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.403119] env[61906]: DEBUG oslo_vmware.api [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356560, 'name': PowerOffVM_Task, 'duration_secs': 0.288791} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.403624] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 805.403624] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 805.403919] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a81a3b20-0034-4f7b-b726-b6b4c571353f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.416504] env[61906]: DEBUG nova.network.neutron [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.466281] env[61906]: DEBUG nova.compute.manager [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Received event network-changed-76876157-42ef-4519-88b2-ca9a600e3b41 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.466566] env[61906]: DEBUG nova.compute.manager [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Refreshing instance network info cache due to event network-changed-76876157-42ef-4519-88b2-ca9a600e3b41. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 805.466797] env[61906]: DEBUG oslo_concurrency.lockutils [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] Acquiring lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.466947] env[61906]: DEBUG oslo_concurrency.lockutils [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] Acquired lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.467140] env[61906]: DEBUG nova.network.neutron [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Refreshing network info cache for port 76876157-42ef-4519-88b2-ca9a600e3b41 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 805.477975] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356561, 'name': Rename_Task, 'duration_secs': 0.14825} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.478335] env[61906]: DEBUG nova.compute.manager [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 805.480893] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 805.481841] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-48e7ad68-5131-44ca-a463-7b24c70b0ef5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.491637] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Waiting for the task: (returnval){ [ 805.491637] env[61906]: value = "task-1356567" [ 805.491637] env[61906]: _type = "Task" [ 805.491637] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.501274] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356567, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.532255] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 805.532425] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 805.532636] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Deleting the datastore file [datastore2] 7943ab1a-82df-48dd-874c-15dc015db51e {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 805.532935] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fa33ada-91e8-40a5-a819-1b89bdd32d4e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.541419] env[61906]: DEBUG oslo_vmware.api [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for the task: (returnval){ [ 805.541419] env[61906]: value = "task-1356568" [ 805.541419] env[61906]: _type = "Task" [ 805.541419] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.550403] env[61906]: DEBUG oslo_vmware.api [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356568, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.740339] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525be344-e1d4-87c6-9e1f-29a52b797b9e, 'name': SearchDatastore_Task, 'duration_secs': 0.009347} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.740622] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.740887] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 9a23071d-40fd-4446-aa03-ecbddcafe3d6/9a23071d-40fd-4446-aa03-ecbddcafe3d6.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 805.741167] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-526ade8a-6e10-4fc4-abc9-00eee6d9a199 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.747312] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 805.747312] env[61906]: value = "task-1356569" [ 805.747312] env[61906]: _type = "Task" [ 805.747312] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.755917] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356569, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.879465] env[61906]: DEBUG oslo_vmware.api [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356564, 'name': PowerOffVM_Task, 'duration_secs': 0.239935} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.880179] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 805.880385] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 805.880644] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06a1eeaa-ab36-4f4e-ba76-f8956db828fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.887722] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356565, 'name': CreateVM_Task, 'duration_secs': 0.319168} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.888585] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 805.889368] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.889538] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.889848] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.890108] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b22430bc-28e4-4607-9cd8-d55a39739291 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.894724] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 805.894724] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a2e07c-f98d-e239-17f9-b0cbcc3cc7d7" [ 805.894724] env[61906]: _type = "Task" [ 805.894724] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.906396] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a2e07c-f98d-e239-17f9-b0cbcc3cc7d7, 'name': SearchDatastore_Task, 'duration_secs': 0.00892} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.906932] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.906932] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 805.907966] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.907966] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.907966] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 805.907966] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-210747b3-0ee2-4e3c-86c8-4f5c91b1e2fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.915036] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 805.915253] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 805.915960] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f04b423-9747-4d1a-8f77-ef2b314c6d30 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.921441] env[61906]: INFO nova.compute.manager [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] [instance: c3cb1a2d-e905-497c-a4de-8c6579c37156] Took 1.02 seconds to deallocate network for instance. [ 805.924213] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 805.924213] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52501803-b6f6-8d69-db94-8b9c59a6c73b" [ 805.924213] env[61906]: _type = "Task" [ 805.924213] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.936610] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52501803-b6f6-8d69-db94-8b9c59a6c73b, 'name': SearchDatastore_Task, 'duration_secs': 0.011078} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.939818] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48fb7318-5fe8-444f-bf0a-0bcb58e85133 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.945482] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 805.945482] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52336c51-d1da-a74f-99c4-69262005bd20" [ 805.945482] env[61906]: _type = "Task" [ 805.945482] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.953961] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52336c51-d1da-a74f-99c4-69262005bd20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.997426] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 805.997635] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 805.997811] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Deleting the datastore file [datastore2] 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 805.999016] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f0f91eed-923a-43db-a4c1-21c76dabd099 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.001254] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.007102] env[61906]: DEBUG oslo_vmware.api [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356567, 'name': PowerOnVM_Task, 'duration_secs': 0.490318} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.010612] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 806.010832] env[61906]: INFO nova.compute.manager [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Took 7.82 seconds to spawn the instance on the hypervisor. [ 806.011018] env[61906]: DEBUG nova.compute.manager [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 806.011830] env[61906]: DEBUG oslo_vmware.api [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Waiting for the task: (returnval){ [ 806.011830] env[61906]: value = "task-1356571" [ 806.011830] env[61906]: _type = "Task" [ 806.011830] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.012756] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9839c204-89aa-48b5-8550-9f9ebb17a5fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.024461] env[61906]: DEBUG oslo_vmware.api [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356571, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.054801] env[61906]: DEBUG oslo_vmware.api [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Task: {'id': task-1356568, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185017} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.058183] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 806.058448] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 806.058690] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 806.058928] env[61906]: INFO nova.compute.manager [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Took 1.18 seconds to destroy the instance on the hypervisor. [ 806.059236] env[61906]: DEBUG oslo.service.loopingcall [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.059682] env[61906]: DEBUG nova.compute.manager [-] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.059774] env[61906]: DEBUG nova.network.neutron [-] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 806.073184] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a3d6fa-e6aa-4a72-b262-83a124f32fbc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.081395] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f4380c5-1509-45d2-b171-a40745c0cc56 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.120361] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3309d90-92c0-4109-9c29-52367553a275 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.128923] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb140da7-a42e-4a00-83fb-27b12175f817 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.143055] env[61906]: DEBUG nova.compute.provider_tree [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.258777] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356569, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.324351] env[61906]: DEBUG nova.network.neutron [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Updated VIF entry in instance network info cache for port 76876157-42ef-4519-88b2-ca9a600e3b41. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 806.324818] env[61906]: DEBUG nova.network.neutron [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Updating instance_info_cache with network_info: [{"id": "76876157-42ef-4519-88b2-ca9a600e3b41", "address": "fa:16:3e:66:6f:62", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76876157-42", "ovs_interfaceid": "76876157-42ef-4519-88b2-ca9a600e3b41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.396535] env[61906]: DEBUG nova.compute.manager [req-e96efdf8-d22d-485e-94dc-ac0ea6896742 req-07ad691e-d8c8-402a-9f05-1b511e6c73cc service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Received event network-vif-deleted-f38d9ad1-e3be-4227-981d-b2b14c6cb912 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.396535] env[61906]: INFO nova.compute.manager [req-e96efdf8-d22d-485e-94dc-ac0ea6896742 req-07ad691e-d8c8-402a-9f05-1b511e6c73cc service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Neutron deleted interface f38d9ad1-e3be-4227-981d-b2b14c6cb912; detaching it from the instance and deleting it from the info cache [ 806.396535] env[61906]: DEBUG nova.network.neutron [req-e96efdf8-d22d-485e-94dc-ac0ea6896742 req-07ad691e-d8c8-402a-9f05-1b511e6c73cc service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.456200] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52336c51-d1da-a74f-99c4-69262005bd20, 'name': SearchDatastore_Task, 'duration_secs': 0.010913} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.456589] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.456858] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 5adaa660-b736-4c11-9141-846cf475ccd5/5adaa660-b736-4c11-9141-846cf475ccd5.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 806.457125] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4eafd833-b6ff-4251-9785-d633aa21c405 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.463706] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 806.463706] env[61906]: value = "task-1356572" [ 806.463706] env[61906]: _type = "Task" [ 806.463706] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.472019] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356572, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.524557] env[61906]: DEBUG oslo_vmware.api [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356571, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.540903] env[61906]: INFO nova.compute.manager [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Took 26.16 seconds to build instance. [ 806.648345] env[61906]: DEBUG nova.scheduler.client.report [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.758363] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356569, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.629351} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.758652] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 9a23071d-40fd-4446-aa03-ecbddcafe3d6/9a23071d-40fd-4446-aa03-ecbddcafe3d6.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 806.758874] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 806.759134] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0b76120-8ad6-44e4-ab80-842bdb7b8c19 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.765262] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 806.765262] env[61906]: value = "task-1356573" [ 806.765262] env[61906]: _type = "Task" [ 806.765262] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.773535] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356573, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.828198] env[61906]: DEBUG oslo_concurrency.lockutils [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] Releasing lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.828567] env[61906]: DEBUG nova.compute.manager [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Received event network-changed-f38d9ad1-e3be-4227-981d-b2b14c6cb912 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.828764] env[61906]: DEBUG nova.compute.manager [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Refreshing instance network info cache due to event network-changed-f38d9ad1-e3be-4227-981d-b2b14c6cb912. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 806.829010] env[61906]: DEBUG oslo_concurrency.lockutils [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] Acquiring lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.829162] env[61906]: DEBUG oslo_concurrency.lockutils [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] Acquired lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.829408] env[61906]: DEBUG nova.network.neutron [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Refreshing network info cache for port f38d9ad1-e3be-4227-981d-b2b14c6cb912 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 806.856019] env[61906]: DEBUG nova.network.neutron [-] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.899154] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0441e950-5e7a-4e0b-82a2-280dd7a7dcd4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.908147] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db50b48-3062-4eb4-b64e-a9e65678f007 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.935784] env[61906]: DEBUG nova.compute.manager [req-e96efdf8-d22d-485e-94dc-ac0ea6896742 req-07ad691e-d8c8-402a-9f05-1b511e6c73cc service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Detach interface failed, port_id=f38d9ad1-e3be-4227-981d-b2b14c6cb912, reason: Instance 7943ab1a-82df-48dd-874c-15dc015db51e could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 806.966168] env[61906]: INFO nova.scheduler.client.report [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Deleted allocations for instance c3cb1a2d-e905-497c-a4de-8c6579c37156 [ 806.981811] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356572, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.026760] env[61906]: DEBUG oslo_vmware.api [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Task: {'id': task-1356571, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.818946} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.027037] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 807.027227] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 807.027416] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 807.027594] env[61906]: INFO nova.compute.manager [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Took 1.69 seconds to destroy the instance on the hypervisor. [ 807.027834] env[61906]: DEBUG oslo.service.loopingcall [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.028037] env[61906]: DEBUG nova.compute.manager [-] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 807.028134] env[61906]: DEBUG nova.network.neutron [-] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 807.043432] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da72d54a-1164-4288-a580-da33e3db9835 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Lock "8ab6b473-d13c-4341-9789-992ac3aba6a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.122s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.153137] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.377s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.153694] env[61906]: DEBUG nova.compute.manager [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 807.157898] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.880s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.159018] env[61906]: INFO nova.compute.claims [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.277375] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356573, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069428} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.277741] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 807.278661] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e6b442-132a-42f7-aede-d091aae034b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.305969] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 9a23071d-40fd-4446-aa03-ecbddcafe3d6/9a23071d-40fd-4446-aa03-ecbddcafe3d6.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.306393] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-598bcc45-5e7c-444a-b83f-241184474039 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.329634] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 807.329634] env[61906]: value = "task-1356574" [ 807.329634] env[61906]: _type = "Task" [ 807.329634] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.342271] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356574, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.356854] env[61906]: DEBUG nova.network.neutron [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.358800] env[61906]: INFO nova.compute.manager [-] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Took 1.30 seconds to deallocate network for instance. [ 807.460523] env[61906]: DEBUG nova.network.neutron [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.477463] env[61906]: DEBUG oslo_concurrency.lockutils [None req-067cbbea-5e83-4dc9-b45e-c60a9015a51f tempest-AttachVolumeTestJSON-246068195 tempest-AttachVolumeTestJSON-246068195-project-member] Lock "c3cb1a2d-e905-497c-a4de-8c6579c37156" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.690s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.483387] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356572, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.546776] env[61906]: DEBUG nova.compute.manager [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 807.658466] env[61906]: DEBUG nova.compute.utils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 807.659787] env[61906]: DEBUG nova.compute.manager [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 807.659960] env[61906]: DEBUG nova.network.neutron [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 807.711990] env[61906]: DEBUG nova.policy [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9dba4398586641c9adfe209ea853538f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17447d31611a46bbabae82a3ea2a2f22', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 807.842372] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356574, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.866484] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.963894] env[61906]: DEBUG oslo_concurrency.lockutils [req-aaf36530-6fff-413a-8420-1a782e30c0c5 req-7c221628-31a4-482e-bb30-88e4846f1c2c service nova] Releasing lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.986827] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356572, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.074873} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.986827] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 5adaa660-b736-4c11-9141-846cf475ccd5/5adaa660-b736-4c11-9141-846cf475ccd5.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 807.986827] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.986827] env[61906]: DEBUG nova.compute.manager [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 807.986827] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7fa11e57-eb22-4d34-9276-611f75d0b27e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.997210] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 807.997210] env[61906]: value = "task-1356575" [ 807.997210] env[61906]: _type = "Task" [ 807.997210] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.998903] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Acquiring lock "8ab6b473-d13c-4341-9789-992ac3aba6a2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.999508] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Lock "8ab6b473-d13c-4341-9789-992ac3aba6a2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.000045] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Acquiring lock "8ab6b473-d13c-4341-9789-992ac3aba6a2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.000747] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Lock "8ab6b473-d13c-4341-9789-992ac3aba6a2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.001063] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Lock "8ab6b473-d13c-4341-9789-992ac3aba6a2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.003543] env[61906]: INFO nova.compute.manager [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Terminating instance [ 808.009131] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356575, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.010096] env[61906]: DEBUG nova.compute.manager [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 808.010257] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 808.011204] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce33fac0-dcee-400d-a423-ec54e782d7ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.019242] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 808.019713] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de9665d0-9671-4295-bef5-a27e937fc1fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.025548] env[61906]: DEBUG oslo_vmware.api [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Waiting for the task: (returnval){ [ 808.025548] env[61906]: value = "task-1356576" [ 808.025548] env[61906]: _type = "Task" [ 808.025548] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.035876] env[61906]: DEBUG oslo_vmware.api [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356576, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.043145] env[61906]: DEBUG nova.network.neutron [-] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.074368] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.094905] env[61906]: DEBUG nova.network.neutron [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Successfully created port: a4af9140-6653-413f-a40e-0b0c021dc68c {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 808.163702] env[61906]: DEBUG nova.compute.manager [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 808.342514] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356574, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.437820] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d605e34-b4b3-49a6-bb9f-b1558c7da3fd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.445531] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf9c6dd-9c74-4d71-a374-c41fdce92bd2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.476680] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3e7aa1-a1e8-4c61-ab2f-e16d6a62e9e0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.485189] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dece0663-828c-469b-9b0b-b394b54dfe42 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.505682] env[61906]: DEBUG nova.compute.manager [req-a6c451b4-7e83-43d6-b26d-b6a393ac3bed req-70ecc30b-7b3a-4676-8874-70cc3895980b service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Received event network-vif-deleted-47fd5cbf-baa2-424e-a003-1d4116844aad {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.505892] env[61906]: DEBUG nova.compute.manager [req-a6c451b4-7e83-43d6-b26d-b6a393ac3bed req-70ecc30b-7b3a-4676-8874-70cc3895980b service nova] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Received event network-vif-deleted-2bfbe252-2380-4a3c-a47d-39615dfa7991 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.506716] env[61906]: DEBUG nova.compute.provider_tree [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.516868] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356575, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068206} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.517406] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.518811] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8f7389-800a-4e6b-a670-ba4b11c1eabc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.522119] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.544627] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 5adaa660-b736-4c11-9141-846cf475ccd5/5adaa660-b736-4c11-9141-846cf475ccd5.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.546605] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5591b5af-a49d-4f6e-a4a9-6eaa5ef73c16 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.561079] env[61906]: INFO nova.compute.manager [-] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Took 1.53 seconds to deallocate network for instance. [ 808.571164] env[61906]: DEBUG oslo_vmware.api [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356576, 'name': PowerOffVM_Task, 'duration_secs': 0.237215} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.572531] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 808.572712] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 808.573033] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 808.573033] env[61906]: value = "task-1356577" [ 808.573033] env[61906]: _type = "Task" [ 808.573033] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.573223] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0af08409-2bef-4d94-8e86-c6f5723f1f4c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.583196] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356577, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.650739] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 808.650961] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 808.651160] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Deleting the datastore file [datastore2] 8ab6b473-d13c-4341-9789-992ac3aba6a2 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 808.651427] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-639df65c-7def-4815-97f2-cccd212dbce3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.663023] env[61906]: DEBUG oslo_vmware.api [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Waiting for the task: (returnval){ [ 808.663023] env[61906]: value = "task-1356579" [ 808.663023] env[61906]: _type = "Task" [ 808.663023] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.667828] env[61906]: DEBUG oslo_vmware.api [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356579, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.844943] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356574, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.009816] env[61906]: DEBUG nova.scheduler.client.report [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.068454] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.087456] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356577, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.171219] env[61906]: DEBUG oslo_vmware.api [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Task: {'id': task-1356579, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.231224} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.171467] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 809.171643] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 809.171814] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 809.171981] env[61906]: INFO nova.compute.manager [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Took 1.16 seconds to destroy the instance on the hypervisor. [ 809.172228] env[61906]: DEBUG oslo.service.loopingcall [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.172419] env[61906]: DEBUG nova.compute.manager [-] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 809.172510] env[61906]: DEBUG nova.network.neutron [-] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 809.178250] env[61906]: DEBUG nova.compute.manager [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 809.205523] env[61906]: DEBUG nova.virt.hardware [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 809.205733] env[61906]: DEBUG nova.virt.hardware [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 809.205893] env[61906]: DEBUG nova.virt.hardware [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.206092] env[61906]: DEBUG nova.virt.hardware [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 809.206273] env[61906]: DEBUG nova.virt.hardware [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.206450] env[61906]: DEBUG nova.virt.hardware [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 809.206659] env[61906]: DEBUG nova.virt.hardware [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 809.206833] env[61906]: DEBUG nova.virt.hardware [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 809.206973] env[61906]: DEBUG nova.virt.hardware [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 809.207598] env[61906]: DEBUG nova.virt.hardware [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 809.207803] env[61906]: DEBUG nova.virt.hardware [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.208784] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a45acb-bb61-456b-8224-90143496e25f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.217291] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524048f5-3ea3-4e34-922f-26f666edd522 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.344587] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356574, 'name': ReconfigVM_Task, 'duration_secs': 1.977654} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.344861] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 9a23071d-40fd-4446-aa03-ecbddcafe3d6/9a23071d-40fd-4446-aa03-ecbddcafe3d6.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.345585] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-832aae89-9c10-402c-b0e2-47da432f4332 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.352115] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 809.352115] env[61906]: value = "task-1356580" [ 809.352115] env[61906]: _type = "Task" [ 809.352115] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.359951] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356580, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.515271] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.515794] env[61906]: DEBUG nova.compute.manager [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.519717] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.349s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.519717] env[61906]: DEBUG nova.objects.instance [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Lazy-loading 'resources' on Instance uuid a6e54432-a58e-49ba-a5c8-e6188f2ea4ff {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 809.585950] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356577, 'name': ReconfigVM_Task, 'duration_secs': 0.96214} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.586464] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 5adaa660-b736-4c11-9141-846cf475ccd5/5adaa660-b736-4c11-9141-846cf475ccd5.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.587103] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7513ab75-9042-4d34-b0f6-a3a342b184ca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.596666] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 809.596666] env[61906]: value = "task-1356581" [ 809.596666] env[61906]: _type = "Task" [ 809.596666] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.608619] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356581, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.664137] env[61906]: DEBUG nova.compute.manager [req-f267409e-83a8-40a1-b740-04f34cbc1352 req-d03bf66d-6668-410c-92c4-b5cabdf8ac35 service nova] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Received event network-vif-deleted-56efd84b-d24d-49e4-870b-7361a46db687 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.664500] env[61906]: INFO nova.compute.manager [req-f267409e-83a8-40a1-b740-04f34cbc1352 req-d03bf66d-6668-410c-92c4-b5cabdf8ac35 service nova] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Neutron deleted interface 56efd84b-d24d-49e4-870b-7361a46db687; detaching it from the instance and deleting it from the info cache [ 809.664773] env[61906]: DEBUG nova.network.neutron [req-f267409e-83a8-40a1-b740-04f34cbc1352 req-d03bf66d-6668-410c-92c4-b5cabdf8ac35 service nova] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.867354] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356580, 'name': Rename_Task, 'duration_secs': 0.140701} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.868646] env[61906]: DEBUG nova.network.neutron [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Successfully updated port: a4af9140-6653-413f-a40e-0b0c021dc68c {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 809.869923] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 809.870216] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e03cd1b-44fa-443b-91d0-ca537b4079da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.878220] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 809.878220] env[61906]: value = "task-1356582" [ 809.878220] env[61906]: _type = "Task" [ 809.878220] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.889705] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356582, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.020773] env[61906]: DEBUG nova.compute.utils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.022770] env[61906]: DEBUG nova.compute.manager [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 810.022770] env[61906]: DEBUG nova.network.neutron [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 810.097092] env[61906]: DEBUG nova.policy [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '231e62b44cd94249ab22ef8c43e60c96', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24249dfdffc046ddb41d0da91203fa4a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.120766] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356581, 'name': Rename_Task, 'duration_secs': 0.130494} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.121224] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 810.121754] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8deaa3ed-5348-4999-8425-c9d40a85665b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.132715] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 810.132715] env[61906]: value = "task-1356583" [ 810.132715] env[61906]: _type = "Task" [ 810.132715] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.143953] env[61906]: DEBUG nova.network.neutron [-] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.153937] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356583, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.173226] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fcdbc364-ecd8-4719-bd97-3a8dda0b9fb8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.184788] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6d351c-8641-438b-ad58-4543aa18a1b9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.212700] env[61906]: DEBUG nova.compute.manager [req-f267409e-83a8-40a1-b740-04f34cbc1352 req-d03bf66d-6668-410c-92c4-b5cabdf8ac35 service nova] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Detach interface failed, port_id=56efd84b-d24d-49e4-870b-7361a46db687, reason: Instance 8ab6b473-d13c-4341-9789-992ac3aba6a2 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 810.363245] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746f26c3-d3fa-48fb-ac0c-62b69dc91354 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.371194] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed34b4f5-7e69-4a4d-984e-4871cf62789d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.377016] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "refresh_cache-51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.377016] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquired lock "refresh_cache-51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.377016] env[61906]: DEBUG nova.network.neutron [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 810.412530] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58bf8846-9b4c-40df-ad76-37e660a1cf8e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.418724] env[61906]: DEBUG oslo_vmware.api [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356582, 'name': PowerOnVM_Task, 'duration_secs': 0.498298} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.422700] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 810.422801] env[61906]: INFO nova.compute.manager [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Took 9.84 seconds to spawn the instance on the hypervisor. [ 810.422978] env[61906]: DEBUG nova.compute.manager [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.424653] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6a0d68-1dab-4466-8872-8fab0bb83686 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.430537] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b5a11c5-c056-48fc-b11d-0305939d15e0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.449355] env[61906]: DEBUG nova.compute.provider_tree [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.452431] env[61906]: DEBUG nova.network.neutron [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Successfully created port: e291dedf-5c3f-4d78-bad3-295b2d1e2f75 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.525984] env[61906]: DEBUG nova.compute.manager [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Received event network-vif-plugged-a4af9140-6653-413f-a40e-0b0c021dc68c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.526297] env[61906]: DEBUG oslo_concurrency.lockutils [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] Acquiring lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.526583] env[61906]: DEBUG oslo_concurrency.lockutils [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.526665] env[61906]: DEBUG oslo_concurrency.lockutils [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.526816] env[61906]: DEBUG nova.compute.manager [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] No waiting events found dispatching network-vif-plugged-a4af9140-6653-413f-a40e-0b0c021dc68c {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 810.526980] env[61906]: WARNING nova.compute.manager [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Received unexpected event network-vif-plugged-a4af9140-6653-413f-a40e-0b0c021dc68c for instance with vm_state building and task_state spawning. [ 810.527153] env[61906]: DEBUG nova.compute.manager [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Received event network-changed-a4af9140-6653-413f-a40e-0b0c021dc68c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.527306] env[61906]: DEBUG nova.compute.manager [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Refreshing instance network info cache due to event network-changed-a4af9140-6653-413f-a40e-0b0c021dc68c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 810.527470] env[61906]: DEBUG oslo_concurrency.lockutils [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] Acquiring lock "refresh_cache-51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.529299] env[61906]: DEBUG nova.compute.manager [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.644467] env[61906]: DEBUG oslo_vmware.api [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356583, 'name': PowerOnVM_Task, 'duration_secs': 0.504869} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.644467] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 810.644654] env[61906]: INFO nova.compute.manager [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Took 7.68 seconds to spawn the instance on the hypervisor. [ 810.644824] env[61906]: DEBUG nova.compute.manager [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 810.645681] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20653f2-f14e-49bb-9174-e8cb57f255dc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.649790] env[61906]: INFO nova.compute.manager [-] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Took 1.48 seconds to deallocate network for instance. [ 810.905662] env[61906]: DEBUG nova.network.neutron [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.948946] env[61906]: INFO nova.compute.manager [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Took 29.94 seconds to build instance. [ 810.957121] env[61906]: DEBUG nova.scheduler.client.report [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.049965] env[61906]: DEBUG nova.network.neutron [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Updating instance_info_cache with network_info: [{"id": "a4af9140-6653-413f-a40e-0b0c021dc68c", "address": "fa:16:3e:c3:05:8a", "network": {"id": "b4ea38e3-d0e0-48a3-9632-772ca8257e37", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-540734152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17447d31611a46bbabae82a3ea2a2f22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90328c7b-15c4-4742-805b-755248d67029", "external-id": "nsx-vlan-transportzone-860", "segmentation_id": 860, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4af9140-66", "ovs_interfaceid": "a4af9140-6653-413f-a40e-0b0c021dc68c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.165952] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.169186] env[61906]: INFO nova.compute.manager [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Took 28.22 seconds to build instance. [ 811.450853] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6b6a5959-b912-4809-bd6f-bfde33d68ff0 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.459s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.461949] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.943s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.465368] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.851s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.466401] env[61906]: INFO nova.compute.claims [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.484618] env[61906]: INFO nova.scheduler.client.report [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Deleted allocations for instance a6e54432-a58e-49ba-a5c8-e6188f2ea4ff [ 811.537932] env[61906]: DEBUG nova.compute.manager [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.552431] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Releasing lock "refresh_cache-51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.552855] env[61906]: DEBUG nova.compute.manager [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Instance network_info: |[{"id": "a4af9140-6653-413f-a40e-0b0c021dc68c", "address": "fa:16:3e:c3:05:8a", "network": {"id": "b4ea38e3-d0e0-48a3-9632-772ca8257e37", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-540734152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17447d31611a46bbabae82a3ea2a2f22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90328c7b-15c4-4742-805b-755248d67029", "external-id": "nsx-vlan-transportzone-860", "segmentation_id": 860, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4af9140-66", "ovs_interfaceid": "a4af9140-6653-413f-a40e-0b0c021dc68c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 811.553183] env[61906]: DEBUG oslo_concurrency.lockutils [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] Acquired lock "refresh_cache-51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.553368] env[61906]: DEBUG nova.network.neutron [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Refreshing network info cache for port a4af9140-6653-413f-a40e-0b0c021dc68c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 811.555081] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:05:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '90328c7b-15c4-4742-805b-755248d67029', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a4af9140-6653-413f-a40e-0b0c021dc68c', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.563055] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Creating folder: Project (17447d31611a46bbabae82a3ea2a2f22). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.566201] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d93f70df-88d0-461c-82a1-48cb5b571c54 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.574707] env[61906]: DEBUG nova.virt.hardware [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.574900] env[61906]: DEBUG nova.virt.hardware [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.575080] env[61906]: DEBUG nova.virt.hardware [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.575269] env[61906]: DEBUG nova.virt.hardware [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.575429] env[61906]: DEBUG nova.virt.hardware [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.575576] env[61906]: DEBUG nova.virt.hardware [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.575779] env[61906]: DEBUG nova.virt.hardware [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.575935] env[61906]: DEBUG nova.virt.hardware [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.576114] env[61906]: DEBUG nova.virt.hardware [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.576278] env[61906]: DEBUG nova.virt.hardware [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.576450] env[61906]: DEBUG nova.virt.hardware [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.577507] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ac377c-069a-46e9-b584-288b9ab9cb97 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.581075] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Created folder: Project (17447d31611a46bbabae82a3ea2a2f22) in parent group-v288914. [ 811.581261] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Creating folder: Instances. Parent ref: group-v288956. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.581784] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-20b9bb04-2422-40cd-84cb-c718dc699e8b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.586925] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406ffb40-67e9-40fe-b3d5-6afd24bfe68b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.591785] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Created folder: Instances in parent group-v288956. [ 811.592016] env[61906]: DEBUG oslo.service.loopingcall [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.592537] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 811.592730] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b64ad6d9-a35c-4700-86eb-67dbf7c2c466 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.618604] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.618604] env[61906]: value = "task-1356586" [ 811.618604] env[61906]: _type = "Task" [ 811.618604] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.625633] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356586, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.671687] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b214f835-9a06-4d7f-8001-7dd81b9f2a61 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "5adaa660-b736-4c11-9141-846cf475ccd5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.585s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.954319] env[61906]: DEBUG nova.compute.manager [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 811.991653] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d8fb14d0-da06-4c7d-8f2a-061ff7620620 tempest-ServerShowV254Test-1839520566 tempest-ServerShowV254Test-1839520566-project-member] Lock "a6e54432-a58e-49ba-a5c8-e6188f2ea4ff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.053s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.097147] env[61906]: DEBUG nova.network.neutron [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Successfully updated port: e291dedf-5c3f-4d78-bad3-295b2d1e2f75 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 812.130445] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356586, 'name': CreateVM_Task, 'duration_secs': 0.31003} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.130657] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 812.131358] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.131564] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.132032] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 812.132404] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0d9bd6d-ef4a-4434-813c-cdabc0df2835 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.136820] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 812.136820] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5252bc59-dd30-0f90-076c-5e655e05a0c8" [ 812.136820] env[61906]: _type = "Task" [ 812.136820] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.144567] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5252bc59-dd30-0f90-076c-5e655e05a0c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.174527] env[61906]: DEBUG nova.compute.manager [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 812.361413] env[61906]: DEBUG oslo_concurrency.lockutils [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.361705] env[61906]: DEBUG oslo_concurrency.lockutils [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.480235] env[61906]: DEBUG oslo_concurrency.lockutils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.555064] env[61906]: DEBUG nova.compute.manager [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Received event network-vif-plugged-e291dedf-5c3f-4d78-bad3-295b2d1e2f75 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.555400] env[61906]: DEBUG oslo_concurrency.lockutils [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] Acquiring lock "79452791-59cb-4722-bb4a-8e59d8c4e641-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.555609] env[61906]: DEBUG oslo_concurrency.lockutils [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] Lock "79452791-59cb-4722-bb4a-8e59d8c4e641-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.555839] env[61906]: DEBUG oslo_concurrency.lockutils [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] Lock "79452791-59cb-4722-bb4a-8e59d8c4e641-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.555940] env[61906]: DEBUG nova.compute.manager [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] No waiting events found dispatching network-vif-plugged-e291dedf-5c3f-4d78-bad3-295b2d1e2f75 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 812.556157] env[61906]: WARNING nova.compute.manager [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Received unexpected event network-vif-plugged-e291dedf-5c3f-4d78-bad3-295b2d1e2f75 for instance with vm_state building and task_state spawning. [ 812.556289] env[61906]: DEBUG nova.compute.manager [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Received event network-changed-e291dedf-5c3f-4d78-bad3-295b2d1e2f75 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.556452] env[61906]: DEBUG nova.compute.manager [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Refreshing instance network info cache due to event network-changed-e291dedf-5c3f-4d78-bad3-295b2d1e2f75. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 812.556652] env[61906]: DEBUG oslo_concurrency.lockutils [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] Acquiring lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.556786] env[61906]: DEBUG oslo_concurrency.lockutils [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] Acquired lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.557954] env[61906]: DEBUG nova.network.neutron [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Refreshing network info cache for port e291dedf-5c3f-4d78-bad3-295b2d1e2f75 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 812.601840] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.649330] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5252bc59-dd30-0f90-076c-5e655e05a0c8, 'name': SearchDatastore_Task, 'duration_secs': 0.013219} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.651876] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.652050] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.652281] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.652427] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.652603] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.653100] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-07cd239e-3f5d-4c27-8542-6e86675d5e4e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.666521] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.666600] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 812.667322] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49a3308d-ac7f-4b76-b5f5-b3bd20b2c4cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.672298] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 812.672298] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522a3a9c-c37b-1b06-7427-f92063ba668b" [ 812.672298] env[61906]: _type = "Task" [ 812.672298] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.683450] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522a3a9c-c37b-1b06-7427-f92063ba668b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.684189] env[61906]: DEBUG nova.network.neutron [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Updated VIF entry in instance network info cache for port a4af9140-6653-413f-a40e-0b0c021dc68c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 812.684778] env[61906]: DEBUG nova.network.neutron [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Updating instance_info_cache with network_info: [{"id": "a4af9140-6653-413f-a40e-0b0c021dc68c", "address": "fa:16:3e:c3:05:8a", "network": {"id": "b4ea38e3-d0e0-48a3-9632-772ca8257e37", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-540734152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17447d31611a46bbabae82a3ea2a2f22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90328c7b-15c4-4742-805b-755248d67029", "external-id": "nsx-vlan-transportzone-860", "segmentation_id": 860, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4af9140-66", "ovs_interfaceid": "a4af9140-6653-413f-a40e-0b0c021dc68c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.704027] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.736219] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5487d665-f1ca-4105-83c0-2478b1dee2b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.745293] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222b68ae-1fa7-4ba4-b65b-19ac43765fb3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.776053] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5733fdf-03c7-4730-9dd9-cd402dfc7e16 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.784044] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20e0924-8bfb-4f55-be6e-d999f7ff582c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.799224] env[61906]: DEBUG nova.compute.provider_tree [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.866027] env[61906]: DEBUG nova.compute.utils [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.872812] env[61906]: DEBUG nova.compute.manager [req-b546b846-eee9-43e9-b6a5-9ab22c37c3d5 req-10a8f9b3-ab47-45f9-a28a-1e8338544ebf service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Received event network-changed-76876157-42ef-4519-88b2-ca9a600e3b41 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.872812] env[61906]: DEBUG nova.compute.manager [req-b546b846-eee9-43e9-b6a5-9ab22c37c3d5 req-10a8f9b3-ab47-45f9-a28a-1e8338544ebf service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Refreshing instance network info cache due to event network-changed-76876157-42ef-4519-88b2-ca9a600e3b41. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 812.872812] env[61906]: DEBUG oslo_concurrency.lockutils [req-b546b846-eee9-43e9-b6a5-9ab22c37c3d5 req-10a8f9b3-ab47-45f9-a28a-1e8338544ebf service nova] Acquiring lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.872812] env[61906]: DEBUG oslo_concurrency.lockutils [req-b546b846-eee9-43e9-b6a5-9ab22c37c3d5 req-10a8f9b3-ab47-45f9-a28a-1e8338544ebf service nova] Acquired lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.872812] env[61906]: DEBUG nova.network.neutron [req-b546b846-eee9-43e9-b6a5-9ab22c37c3d5 req-10a8f9b3-ab47-45f9-a28a-1e8338544ebf service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Refreshing network info cache for port 76876157-42ef-4519-88b2-ca9a600e3b41 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 813.105163] env[61906]: DEBUG nova.network.neutron [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 813.185377] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522a3a9c-c37b-1b06-7427-f92063ba668b, 'name': SearchDatastore_Task, 'duration_secs': 0.01027} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.186467] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c42a5d3-91e6-4352-81f3-11f50b210d1c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.189380] env[61906]: DEBUG oslo_concurrency.lockutils [req-198fc7ed-9f82-4ee3-ad10-fc478174b5b2 req-4a416dbb-4e91-43d4-85c6-95a47a919047 service nova] Releasing lock "refresh_cache-51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.193486] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 813.193486] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52340c33-5d70-f2e1-de84-dce3a11dea89" [ 813.193486] env[61906]: _type = "Task" [ 813.193486] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.202135] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52340c33-5d70-f2e1-de84-dce3a11dea89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.216587] env[61906]: DEBUG nova.network.neutron [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.305020] env[61906]: DEBUG nova.scheduler.client.report [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.368531] env[61906]: DEBUG oslo_concurrency.lockutils [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.711617] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52340c33-5d70-f2e1-de84-dce3a11dea89, 'name': SearchDatastore_Task, 'duration_secs': 0.010279} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.711617] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.711617] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83/51fa46a9-b1b3-4115-b1d7-cc4f672f2e83.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 813.711617] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f5abaf8-fa3b-4082-936c-7556ba63e2b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.720198] env[61906]: DEBUG oslo_concurrency.lockutils [req-e0c5d8d7-3300-4452-9db3-1ab04f6de667 req-9f61f8b1-f1ef-42bb-9096-c9c1ba71a9d6 service nova] Releasing lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.720758] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 813.720758] env[61906]: value = "task-1356587" [ 813.720758] env[61906]: _type = "Task" [ 813.720758] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.721425] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.721688] env[61906]: DEBUG nova.network.neutron [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 813.735574] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356587, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.788497] env[61906]: DEBUG nova.network.neutron [req-b546b846-eee9-43e9-b6a5-9ab22c37c3d5 req-10a8f9b3-ab47-45f9-a28a-1e8338544ebf service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Updated VIF entry in instance network info cache for port 76876157-42ef-4519-88b2-ca9a600e3b41. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 813.788866] env[61906]: DEBUG nova.network.neutron [req-b546b846-eee9-43e9-b6a5-9ab22c37c3d5 req-10a8f9b3-ab47-45f9-a28a-1e8338544ebf service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Updating instance_info_cache with network_info: [{"id": "76876157-42ef-4519-88b2-ca9a600e3b41", "address": "fa:16:3e:66:6f:62", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76876157-42", "ovs_interfaceid": "76876157-42ef-4519-88b2-ca9a600e3b41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.810041] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.810300] env[61906]: DEBUG nova.compute.manager [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 813.813051] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.174s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.813278] env[61906]: DEBUG nova.objects.instance [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Lazy-loading 'resources' on Instance uuid 77791983-cef1-4f07-9b62-d52d335b08c8 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 814.248750] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356587, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.268727] env[61906]: DEBUG nova.network.neutron [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.292191] env[61906]: DEBUG oslo_concurrency.lockutils [req-b546b846-eee9-43e9-b6a5-9ab22c37c3d5 req-10a8f9b3-ab47-45f9-a28a-1e8338544ebf service nova] Releasing lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.320587] env[61906]: DEBUG nova.compute.utils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 814.322369] env[61906]: DEBUG nova.compute.manager [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 814.322606] env[61906]: DEBUG nova.network.neutron [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 814.438101] env[61906]: DEBUG oslo_concurrency.lockutils [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.438384] env[61906]: DEBUG oslo_concurrency.lockutils [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.438690] env[61906]: INFO nova.compute.manager [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Attaching volume 96c5a40a-fe62-4fbf-8378-5fa3c290af1a to /dev/sdb [ 814.491644] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf97dc6-03d9-4cbc-9d28-6d18cc90a852 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.504193] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d72a1f4-c8b3-43df-a10a-e24c899702be {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.522724] env[61906]: DEBUG nova.virt.block_device [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Updating existing volume attachment record: 1ef08383-1d59-4258-8b84-dd11dd2f7dc5 {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 814.553468] env[61906]: DEBUG nova.policy [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '231e62b44cd94249ab22ef8c43e60c96', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24249dfdffc046ddb41d0da91203fa4a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 814.558134] env[61906]: DEBUG nova.network.neutron [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Updating instance_info_cache with network_info: [{"id": "e291dedf-5c3f-4d78-bad3-295b2d1e2f75", "address": "fa:16:3e:c8:8d:ef", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape291dedf-5c", "ovs_interfaceid": "e291dedf-5c3f-4d78-bad3-295b2d1e2f75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.640312] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d2f49bf-ecda-486d-a3ca-bf026497531e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.647563] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e783aaac-5ea3-43ef-b374-592f340e1bfb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.684729] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc43b18-5c48-461b-aa90-b6d44e8803a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.692594] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023409ab-501a-4b9f-9e8e-b1abf2700870 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.714936] env[61906]: DEBUG nova.compute.provider_tree [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.734923] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356587, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.90274} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.735559] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83/51fa46a9-b1b3-4115-b1d7-cc4f672f2e83.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 814.735559] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 814.735731] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dec73e84-8d8c-4d4b-97d0-8197d081a897 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.742237] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 814.742237] env[61906]: value = "task-1356590" [ 814.742237] env[61906]: _type = "Task" [ 814.742237] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.751344] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356590, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.826235] env[61906]: DEBUG nova.compute.manager [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 814.893997] env[61906]: DEBUG nova.network.neutron [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Successfully created port: 4c806e0c-9874-4943-9a71-05fc2c40045e {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 815.065525] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.065878] env[61906]: DEBUG nova.compute.manager [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Instance network_info: |[{"id": "e291dedf-5c3f-4d78-bad3-295b2d1e2f75", "address": "fa:16:3e:c8:8d:ef", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape291dedf-5c", "ovs_interfaceid": "e291dedf-5c3f-4d78-bad3-295b2d1e2f75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 815.066348] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:8d:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '680cb499-2a47-482b-af0d-112016ac0e17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e291dedf-5c3f-4d78-bad3-295b2d1e2f75', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 815.073706] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Creating folder: Project (24249dfdffc046ddb41d0da91203fa4a). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.073983] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ba2e5620-a22b-4592-8d2e-3f93f4fe3b30 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.085948] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Created folder: Project (24249dfdffc046ddb41d0da91203fa4a) in parent group-v288914. [ 815.086079] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Creating folder: Instances. Parent ref: group-v288961. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.086356] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35a36e62-578d-4f8c-80a7-4c3c9339cd13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.097249] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Created folder: Instances in parent group-v288961. [ 815.097249] env[61906]: DEBUG oslo.service.loopingcall [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.097249] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 815.097249] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48df67d2-1832-4fbf-94ca-563bb5ff8f32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.115753] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.115753] env[61906]: value = "task-1356594" [ 815.115753] env[61906]: _type = "Task" [ 815.115753] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.123308] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356594, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.218164] env[61906]: DEBUG nova.scheduler.client.report [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.252535] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356590, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.188081} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.252797] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 815.253638] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf87b5e2-469a-4542-b83d-a0b0090bfc3a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.274602] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83/51fa46a9-b1b3-4115-b1d7-cc4f672f2e83.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 815.275100] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88da5fe8-0827-408b-a010-22f4157a8cc5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.293885] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 815.293885] env[61906]: value = "task-1356595" [ 815.293885] env[61906]: _type = "Task" [ 815.293885] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.301600] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356595, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.625099] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356594, 'name': CreateVM_Task, 'duration_secs': 0.376654} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.625374] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 815.626083] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.626270] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.626568] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 815.626812] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5c936a9-aebc-4b2e-9b30-86e944d407db {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.631095] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 815.631095] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52abcd98-3852-243d-11ad-56c02125eedf" [ 815.631095] env[61906]: _type = "Task" [ 815.631095] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.638551] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52abcd98-3852-243d-11ad-56c02125eedf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.723672] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.911s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.726212] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.793s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.727835] env[61906]: INFO nova.compute.claims [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 815.746244] env[61906]: INFO nova.scheduler.client.report [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Deleted allocations for instance 77791983-cef1-4f07-9b62-d52d335b08c8 [ 815.804312] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356595, 'name': ReconfigVM_Task, 'duration_secs': 0.28852} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.805502] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83/51fa46a9-b1b3-4115-b1d7-cc4f672f2e83.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.806238] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52e826a7-b182-4eca-b608-24f4ead04f5a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.813096] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 815.813096] env[61906]: value = "task-1356596" [ 815.813096] env[61906]: _type = "Task" [ 815.813096] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.824918] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356596, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.836704] env[61906]: DEBUG nova.compute.manager [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 815.863115] env[61906]: DEBUG nova.virt.hardware [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 815.863385] env[61906]: DEBUG nova.virt.hardware [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 815.863546] env[61906]: DEBUG nova.virt.hardware [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 815.863727] env[61906]: DEBUG nova.virt.hardware [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 815.863869] env[61906]: DEBUG nova.virt.hardware [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 815.864086] env[61906]: DEBUG nova.virt.hardware [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 815.864345] env[61906]: DEBUG nova.virt.hardware [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 815.864509] env[61906]: DEBUG nova.virt.hardware [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 815.864675] env[61906]: DEBUG nova.virt.hardware [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 815.864836] env[61906]: DEBUG nova.virt.hardware [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 815.865010] env[61906]: DEBUG nova.virt.hardware [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 815.866221] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f36f087-8b3d-40f4-86a9-b162137fe2a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.874169] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e2e9b9-e405-4f96-9a20-8469c0a2a6e2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.140811] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52abcd98-3852-243d-11ad-56c02125eedf, 'name': SearchDatastore_Task, 'duration_secs': 0.031076} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.141129] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.141363] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 816.141931] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.141931] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.141931] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 816.142175] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7a4df7c1-0307-4f30-906d-79267daf7058 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.150343] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 816.150551] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 816.151348] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae720e22-a6e2-437f-bdd6-55a0c7853ba9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.156691] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 816.156691] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a9b8e0-c765-e827-8455-2ee696165332" [ 816.156691] env[61906]: _type = "Task" [ 816.156691] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.164664] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a9b8e0-c765-e827-8455-2ee696165332, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.256281] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d37c10a-d266-46b9-b1c8-36952ce15bfc tempest-InstanceActionsNegativeTestJSON-1210204371 tempest-InstanceActionsNegativeTestJSON-1210204371-project-member] Lock "77791983-cef1-4f07-9b62-d52d335b08c8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.448s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.324674] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356596, 'name': Rename_Task, 'duration_secs': 0.131696} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.324953] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 816.328593] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-72f636e7-032e-4412-9f67-f6c15b049eba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.338618] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 816.338618] env[61906]: value = "task-1356597" [ 816.338618] env[61906]: _type = "Task" [ 816.338618] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.350989] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356597, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.465054] env[61906]: DEBUG nova.compute.manager [req-ee6cdff8-b7ec-4763-b5cd-796babf75a3b req-f16e9f04-f10d-401d-907d-fbb5de6e74f4 service nova] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Received event network-vif-plugged-4c806e0c-9874-4943-9a71-05fc2c40045e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.465356] env[61906]: DEBUG oslo_concurrency.lockutils [req-ee6cdff8-b7ec-4763-b5cd-796babf75a3b req-f16e9f04-f10d-401d-907d-fbb5de6e74f4 service nova] Acquiring lock "7bae658e-b050-4639-b34e-c2671ef5c773-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.465580] env[61906]: DEBUG oslo_concurrency.lockutils [req-ee6cdff8-b7ec-4763-b5cd-796babf75a3b req-f16e9f04-f10d-401d-907d-fbb5de6e74f4 service nova] Lock "7bae658e-b050-4639-b34e-c2671ef5c773-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.465784] env[61906]: DEBUG oslo_concurrency.lockutils [req-ee6cdff8-b7ec-4763-b5cd-796babf75a3b req-f16e9f04-f10d-401d-907d-fbb5de6e74f4 service nova] Lock "7bae658e-b050-4639-b34e-c2671ef5c773-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.465971] env[61906]: DEBUG nova.compute.manager [req-ee6cdff8-b7ec-4763-b5cd-796babf75a3b req-f16e9f04-f10d-401d-907d-fbb5de6e74f4 service nova] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] No waiting events found dispatching network-vif-plugged-4c806e0c-9874-4943-9a71-05fc2c40045e {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 816.466191] env[61906]: WARNING nova.compute.manager [req-ee6cdff8-b7ec-4763-b5cd-796babf75a3b req-f16e9f04-f10d-401d-907d-fbb5de6e74f4 service nova] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Received unexpected event network-vif-plugged-4c806e0c-9874-4943-9a71-05fc2c40045e for instance with vm_state building and task_state spawning. [ 816.666552] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a9b8e0-c765-e827-8455-2ee696165332, 'name': SearchDatastore_Task, 'duration_secs': 0.012855} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.667351] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6f3c8ab-021a-420e-a18d-2412ea8d5c7a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.672364] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 816.672364] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52e73991-c384-280a-5dda-7bdf39366c13" [ 816.672364] env[61906]: _type = "Task" [ 816.672364] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.681194] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52e73991-c384-280a-5dda-7bdf39366c13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.851605] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356597, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.888823] env[61906]: DEBUG nova.network.neutron [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Successfully updated port: 4c806e0c-9874-4943-9a71-05fc2c40045e {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 816.945690] env[61906]: DEBUG nova.compute.manager [req-983724f0-8123-46f4-923f-def3c9093130 req-66d0497a-2037-41d8-ac02-cfeb0e890dec service nova] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Received event network-changed-4c806e0c-9874-4943-9a71-05fc2c40045e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.945889] env[61906]: DEBUG nova.compute.manager [req-983724f0-8123-46f4-923f-def3c9093130 req-66d0497a-2037-41d8-ac02-cfeb0e890dec service nova] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Refreshing instance network info cache due to event network-changed-4c806e0c-9874-4943-9a71-05fc2c40045e. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 816.946124] env[61906]: DEBUG oslo_concurrency.lockutils [req-983724f0-8123-46f4-923f-def3c9093130 req-66d0497a-2037-41d8-ac02-cfeb0e890dec service nova] Acquiring lock "refresh_cache-7bae658e-b050-4639-b34e-c2671ef5c773" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.946305] env[61906]: DEBUG oslo_concurrency.lockutils [req-983724f0-8123-46f4-923f-def3c9093130 req-66d0497a-2037-41d8-ac02-cfeb0e890dec service nova] Acquired lock "refresh_cache-7bae658e-b050-4639-b34e-c2671ef5c773" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.946482] env[61906]: DEBUG nova.network.neutron [req-983724f0-8123-46f4-923f-def3c9093130 req-66d0497a-2037-41d8-ac02-cfeb0e890dec service nova] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Refreshing network info cache for port 4c806e0c-9874-4943-9a71-05fc2c40045e {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 816.962895] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a0cd10-55e8-4474-a902-31d64dc0ce09 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.971255] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeef6cfe-75ea-49f4-86ea-f277c05d01d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.002127] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3f6f67-1ca9-4b27-8a90-13eca30aa496 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.009707] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8445c204-d59f-4b8a-994c-1f0bfbdf6806 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.022630] env[61906]: DEBUG nova.compute.provider_tree [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.182954] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52e73991-c384-280a-5dda-7bdf39366c13, 'name': SearchDatastore_Task, 'duration_secs': 0.014965} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.183273] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.183595] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 79452791-59cb-4722-bb4a-8e59d8c4e641/79452791-59cb-4722-bb4a-8e59d8c4e641.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 817.183789] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1a9a1712-6904-42a4-b364-24fe82e73668 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.193194] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 817.193194] env[61906]: value = "task-1356599" [ 817.193194] env[61906]: _type = "Task" [ 817.193194] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.201107] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356599, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.350389] env[61906]: DEBUG oslo_vmware.api [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356597, 'name': PowerOnVM_Task, 'duration_secs': 0.561071} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.350646] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 817.350848] env[61906]: INFO nova.compute.manager [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Took 8.17 seconds to spawn the instance on the hypervisor. [ 817.351035] env[61906]: DEBUG nova.compute.manager [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 817.351795] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37f3836-976e-4e67-9401-8e6fef6f44d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.392253] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "refresh_cache-7bae658e-b050-4639-b34e-c2671ef5c773" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.480919] env[61906]: DEBUG nova.network.neutron [req-983724f0-8123-46f4-923f-def3c9093130 req-66d0497a-2037-41d8-ac02-cfeb0e890dec service nova] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.525761] env[61906]: DEBUG nova.scheduler.client.report [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.598143] env[61906]: DEBUG nova.network.neutron [req-983724f0-8123-46f4-923f-def3c9093130 req-66d0497a-2037-41d8-ac02-cfeb0e890dec service nova] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.703132] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356599, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.872938] env[61906]: INFO nova.compute.manager [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Took 25.46 seconds to build instance. [ 818.034364] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.034869] env[61906]: DEBUG nova.compute.manager [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 818.037697] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.455s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.040072] env[61906]: INFO nova.compute.claims [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 818.101484] env[61906]: DEBUG oslo_concurrency.lockutils [req-983724f0-8123-46f4-923f-def3c9093130 req-66d0497a-2037-41d8-ac02-cfeb0e890dec service nova] Releasing lock "refresh_cache-7bae658e-b050-4639-b34e-c2671ef5c773" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.101860] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "refresh_cache-7bae658e-b050-4639-b34e-c2671ef5c773" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.102032] env[61906]: DEBUG nova.network.neutron [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 818.204838] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356599, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.563091} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.205391] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 79452791-59cb-4722-bb4a-8e59d8c4e641/79452791-59cb-4722-bb4a-8e59d8c4e641.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 818.205696] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.206014] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-58032442-7b3f-4a3d-8b61-fa8f87614144 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.214463] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 818.214463] env[61906]: value = "task-1356600" [ 818.214463] env[61906]: _type = "Task" [ 818.214463] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.223474] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356600, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.374983] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bbabe1d3-9de0-4e83-a332-1a6adf071ddc tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.563s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.551043] env[61906]: DEBUG nova.compute.utils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 818.555381] env[61906]: DEBUG nova.compute.manager [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 818.555381] env[61906]: DEBUG nova.network.neutron [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 818.602704] env[61906]: DEBUG nova.policy [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ba1658b4ade413b98834102c39a255c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c6414fa082f49c2a568a772a256a8e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 818.699415] env[61906]: DEBUG nova.network.neutron [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.725430] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356600, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093572} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.725752] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 818.726564] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3eac62-3297-4bd1-9ab1-3ae27879ced7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.750937] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 79452791-59cb-4722-bb4a-8e59d8c4e641/79452791-59cb-4722-bb4a-8e59d8c4e641.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 818.751318] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4fd30825-8d87-4530-b554-b5fb9cbca8e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.771595] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 818.771595] env[61906]: value = "task-1356601" [ 818.771595] env[61906]: _type = "Task" [ 818.771595] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.778756] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356601, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.877575] env[61906]: DEBUG nova.compute.manager [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 818.958351] env[61906]: DEBUG nova.network.neutron [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Successfully created port: d803c8e8-7145-412c-aa3e-2545666742eb {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.973839] env[61906]: DEBUG nova.network.neutron [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Updating instance_info_cache with network_info: [{"id": "4c806e0c-9874-4943-9a71-05fc2c40045e", "address": "fa:16:3e:55:c4:86", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c806e0c-98", "ovs_interfaceid": "4c806e0c-9874-4943-9a71-05fc2c40045e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.058588] env[61906]: DEBUG nova.compute.manager [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 819.081234] env[61906]: DEBUG nova.compute.manager [req-14b0c1d9-6069-4689-94c0-00426e9f386a req-ed2ead08-c230-4e42-8266-cae0d2dc3548 service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Received event network-changed-a4af9140-6653-413f-a40e-0b0c021dc68c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.081375] env[61906]: DEBUG nova.compute.manager [req-14b0c1d9-6069-4689-94c0-00426e9f386a req-ed2ead08-c230-4e42-8266-cae0d2dc3548 service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Refreshing instance network info cache due to event network-changed-a4af9140-6653-413f-a40e-0b0c021dc68c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 819.081594] env[61906]: DEBUG oslo_concurrency.lockutils [req-14b0c1d9-6069-4689-94c0-00426e9f386a req-ed2ead08-c230-4e42-8266-cae0d2dc3548 service nova] Acquiring lock "refresh_cache-51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.083173] env[61906]: DEBUG oslo_concurrency.lockutils [req-14b0c1d9-6069-4689-94c0-00426e9f386a req-ed2ead08-c230-4e42-8266-cae0d2dc3548 service nova] Acquired lock "refresh_cache-51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.083173] env[61906]: DEBUG nova.network.neutron [req-14b0c1d9-6069-4689-94c0-00426e9f386a req-ed2ead08-c230-4e42-8266-cae0d2dc3548 service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Refreshing network info cache for port a4af9140-6653-413f-a40e-0b0c021dc68c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 819.096827] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 819.097067] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288960', 'volume_id': '96c5a40a-fe62-4fbf-8378-5fa3c290af1a', 'name': 'volume-96c5a40a-fe62-4fbf-8378-5fa3c290af1a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9a23071d-40fd-4446-aa03-ecbddcafe3d6', 'attached_at': '', 'detached_at': '', 'volume_id': '96c5a40a-fe62-4fbf-8378-5fa3c290af1a', 'serial': '96c5a40a-fe62-4fbf-8378-5fa3c290af1a'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 819.098344] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956e2e4f-bc7d-4033-932d-32be57219e9b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.121966] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2ce8bf-c120-4e90-897c-c25cb3c12bbc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.148315] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] volume-96c5a40a-fe62-4fbf-8378-5fa3c290af1a/volume-96c5a40a-fe62-4fbf-8378-5fa3c290af1a.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.151698] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87c1cc31-c38b-4692-98dd-0bf3a6ed6166 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.171722] env[61906]: DEBUG oslo_vmware.api [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 819.171722] env[61906]: value = "task-1356602" [ 819.171722] env[61906]: _type = "Task" [ 819.171722] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.180610] env[61906]: DEBUG oslo_vmware.api [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356602, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.291018] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356601, 'name': ReconfigVM_Task, 'duration_secs': 0.288202} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.291018] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 79452791-59cb-4722-bb4a-8e59d8c4e641/79452791-59cb-4722-bb4a-8e59d8c4e641.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.291018] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c4e7b7e7-8416-4d74-a302-56207ab89455 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.296991] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 819.296991] env[61906]: value = "task-1356603" [ 819.296991] env[61906]: _type = "Task" [ 819.296991] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.311315] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356603, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.397392] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49cf8a27-c9ab-4594-9c86-29f3bfa8110b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.405415] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-020e06a9-fc3a-4593-b7a8-8b3b80265b83 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.412948] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.439377] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6e9afb-4fa0-4f70-b1ec-b9ce643bd2b1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.448370] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5349ea-c8bd-4a1a-bd7f-cc48e9b97f42 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.460849] env[61906]: DEBUG nova.compute.provider_tree [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.476716] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "refresh_cache-7bae658e-b050-4639-b34e-c2671ef5c773" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.477037] env[61906]: DEBUG nova.compute.manager [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Instance network_info: |[{"id": "4c806e0c-9874-4943-9a71-05fc2c40045e", "address": "fa:16:3e:55:c4:86", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c806e0c-98", "ovs_interfaceid": "4c806e0c-9874-4943-9a71-05fc2c40045e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 819.477445] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:c4:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '680cb499-2a47-482b-af0d-112016ac0e17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c806e0c-9874-4943-9a71-05fc2c40045e', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 819.484932] env[61906]: DEBUG oslo.service.loopingcall [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.485562] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 819.485824] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-151d4378-27e8-4a85-9284-d85f95289219 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.506727] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 819.506727] env[61906]: value = "task-1356604" [ 819.506727] env[61906]: _type = "Task" [ 819.506727] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.514463] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356604, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.684092] env[61906]: DEBUG oslo_vmware.api [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356602, 'name': ReconfigVM_Task, 'duration_secs': 0.429902} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.684876] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Reconfigured VM instance instance-00000041 to attach disk [datastore2] volume-96c5a40a-fe62-4fbf-8378-5fa3c290af1a/volume-96c5a40a-fe62-4fbf-8378-5fa3c290af1a.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.689350] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d98f91c1-33a0-455b-b8c9-8b1b508165a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.709644] env[61906]: DEBUG oslo_vmware.api [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 819.709644] env[61906]: value = "task-1356605" [ 819.709644] env[61906]: _type = "Task" [ 819.709644] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.721944] env[61906]: DEBUG oslo_vmware.api [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356605, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.807180] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356603, 'name': Rename_Task, 'duration_secs': 0.1364} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.809598] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 819.809854] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d4649b7c-be11-4087-9979-3815f8f25bab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.816357] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 819.816357] env[61906]: value = "task-1356606" [ 819.816357] env[61906]: _type = "Task" [ 819.816357] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.824270] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356606, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.853855] env[61906]: DEBUG nova.network.neutron [req-14b0c1d9-6069-4689-94c0-00426e9f386a req-ed2ead08-c230-4e42-8266-cae0d2dc3548 service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Updated VIF entry in instance network info cache for port a4af9140-6653-413f-a40e-0b0c021dc68c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 819.853855] env[61906]: DEBUG nova.network.neutron [req-14b0c1d9-6069-4689-94c0-00426e9f386a req-ed2ead08-c230-4e42-8266-cae0d2dc3548 service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Updating instance_info_cache with network_info: [{"id": "a4af9140-6653-413f-a40e-0b0c021dc68c", "address": "fa:16:3e:c3:05:8a", "network": {"id": "b4ea38e3-d0e0-48a3-9632-772ca8257e37", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-540734152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17447d31611a46bbabae82a3ea2a2f22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90328c7b-15c4-4742-805b-755248d67029", "external-id": "nsx-vlan-transportzone-860", "segmentation_id": 860, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4af9140-66", "ovs_interfaceid": "a4af9140-6653-413f-a40e-0b0c021dc68c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.964160] env[61906]: DEBUG nova.scheduler.client.report [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.016954] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356604, 'name': CreateVM_Task, 'duration_secs': 0.373751} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.017198] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 820.017963] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.018191] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.018550] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 820.018833] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fbdae80-729b-45df-a3dc-1e2a8c5c1ed0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.023440] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 820.023440] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ff9b32-b919-0990-61a3-21584c69c8bd" [ 820.023440] env[61906]: _type = "Task" [ 820.023440] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.031509] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ff9b32-b919-0990-61a3-21584c69c8bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.068702] env[61906]: DEBUG nova.compute.manager [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 820.094874] env[61906]: DEBUG nova.virt.hardware [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.095202] env[61906]: DEBUG nova.virt.hardware [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.095444] env[61906]: DEBUG nova.virt.hardware [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.095631] env[61906]: DEBUG nova.virt.hardware [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.095780] env[61906]: DEBUG nova.virt.hardware [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.095929] env[61906]: DEBUG nova.virt.hardware [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.096156] env[61906]: DEBUG nova.virt.hardware [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.096315] env[61906]: DEBUG nova.virt.hardware [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.096481] env[61906]: DEBUG nova.virt.hardware [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.096645] env[61906]: DEBUG nova.virt.hardware [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.096814] env[61906]: DEBUG nova.virt.hardware [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.097885] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2946cf82-3e89-46fd-a035-6621d0981c3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.105929] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d313da57-b3cb-40f9-a426-b33e01088938 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.219721] env[61906]: DEBUG oslo_vmware.api [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356605, 'name': ReconfigVM_Task, 'duration_secs': 0.165378} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.220112] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288960', 'volume_id': '96c5a40a-fe62-4fbf-8378-5fa3c290af1a', 'name': 'volume-96c5a40a-fe62-4fbf-8378-5fa3c290af1a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9a23071d-40fd-4446-aa03-ecbddcafe3d6', 'attached_at': '', 'detached_at': '', 'volume_id': '96c5a40a-fe62-4fbf-8378-5fa3c290af1a', 'serial': '96c5a40a-fe62-4fbf-8378-5fa3c290af1a'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 820.327279] env[61906]: DEBUG oslo_vmware.api [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356606, 'name': PowerOnVM_Task, 'duration_secs': 0.507288} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.327423] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 820.327653] env[61906]: INFO nova.compute.manager [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Took 8.79 seconds to spawn the instance on the hypervisor. [ 820.327826] env[61906]: DEBUG nova.compute.manager [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.328780] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d0c8d55-4473-47f1-b0ff-a641b0c2eefb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.355955] env[61906]: DEBUG oslo_concurrency.lockutils [req-14b0c1d9-6069-4689-94c0-00426e9f386a req-ed2ead08-c230-4e42-8266-cae0d2dc3548 service nova] Releasing lock "refresh_cache-51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.470250] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.470783] env[61906]: DEBUG nova.compute.manager [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 820.474206] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.473s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.475731] env[61906]: INFO nova.compute.claims [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 820.536665] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ff9b32-b919-0990-61a3-21584c69c8bd, 'name': SearchDatastore_Task, 'duration_secs': 0.02328} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.536999] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.537245] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 820.537476] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.537631] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.537798] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 820.538074] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d42a59d-99c2-4e74-a24e-ac6ea32a4bcf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.546623] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 820.546623] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 820.548415] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e76e8cd-b63b-4b5f-8554-8dd6f9e8ce06 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.553176] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 820.553176] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5227f9a4-c594-5f96-10fc-f012b39a987d" [ 820.553176] env[61906]: _type = "Task" [ 820.553176] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.560659] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5227f9a4-c594-5f96-10fc-f012b39a987d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.606744] env[61906]: DEBUG nova.compute.manager [req-7b2066fe-07e8-48fc-881b-135e21e4f62c req-623ba466-f653-4182-8462-c34c84aaba93 service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Received event network-vif-plugged-d803c8e8-7145-412c-aa3e-2545666742eb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.606744] env[61906]: DEBUG oslo_concurrency.lockutils [req-7b2066fe-07e8-48fc-881b-135e21e4f62c req-623ba466-f653-4182-8462-c34c84aaba93 service nova] Acquiring lock "e133605d-c630-4b9e-a314-bf496c853710-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.606744] env[61906]: DEBUG oslo_concurrency.lockutils [req-7b2066fe-07e8-48fc-881b-135e21e4f62c req-623ba466-f653-4182-8462-c34c84aaba93 service nova] Lock "e133605d-c630-4b9e-a314-bf496c853710-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.606744] env[61906]: DEBUG oslo_concurrency.lockutils [req-7b2066fe-07e8-48fc-881b-135e21e4f62c req-623ba466-f653-4182-8462-c34c84aaba93 service nova] Lock "e133605d-c630-4b9e-a314-bf496c853710-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.606744] env[61906]: DEBUG nova.compute.manager [req-7b2066fe-07e8-48fc-881b-135e21e4f62c req-623ba466-f653-4182-8462-c34c84aaba93 service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] No waiting events found dispatching network-vif-plugged-d803c8e8-7145-412c-aa3e-2545666742eb {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.606744] env[61906]: WARNING nova.compute.manager [req-7b2066fe-07e8-48fc-881b-135e21e4f62c req-623ba466-f653-4182-8462-c34c84aaba93 service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Received unexpected event network-vif-plugged-d803c8e8-7145-412c-aa3e-2545666742eb for instance with vm_state building and task_state spawning. [ 820.695630] env[61906]: DEBUG nova.network.neutron [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Successfully updated port: d803c8e8-7145-412c-aa3e-2545666742eb {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 820.846761] env[61906]: INFO nova.compute.manager [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Took 25.60 seconds to build instance. [ 820.985034] env[61906]: DEBUG nova.compute.utils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 820.987320] env[61906]: DEBUG nova.compute.manager [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 820.987320] env[61906]: DEBUG nova.network.neutron [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 821.039511] env[61906]: DEBUG nova.policy [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd889e15e62b43dea36e5648ca5ad999', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ec67c1b35414992a70b6dfa372ac12e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 821.063854] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5227f9a4-c594-5f96-10fc-f012b39a987d, 'name': SearchDatastore_Task, 'duration_secs': 0.008477} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.065539] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad09f4fc-b71f-409f-b0ac-93c97f17a73c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.070525] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 821.070525] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5239e032-40cf-94b3-815c-42ece7122f54" [ 821.070525] env[61906]: _type = "Task" [ 821.070525] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.079797] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5239e032-40cf-94b3-815c-42ece7122f54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.199886] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.199886] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.199886] env[61906]: DEBUG nova.network.neutron [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 821.260858] env[61906]: DEBUG nova.objects.instance [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lazy-loading 'flavor' on Instance uuid 9a23071d-40fd-4446-aa03-ecbddcafe3d6 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.326657] env[61906]: DEBUG nova.network.neutron [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Successfully created port: 96711ac0-bf70-468e-9ecc-70a5313e4bee {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 821.350315] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fa26a8a2-8787-4224-bdf4-345a43d9e4cc tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "79452791-59cb-4722-bb4a-8e59d8c4e641" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.677s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.490918] env[61906]: DEBUG nova.compute.manager [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 821.585137] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5239e032-40cf-94b3-815c-42ece7122f54, 'name': SearchDatastore_Task, 'duration_secs': 0.012382} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.587668] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.588309] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 7bae658e-b050-4639-b34e-c2671ef5c773/7bae658e-b050-4639-b34e-c2671ef5c773.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 821.588673] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21b6134c-98d9-4351-afd5-4d56e6af9f2f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.595949] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 821.595949] env[61906]: value = "task-1356607" [ 821.595949] env[61906]: _type = "Task" [ 821.595949] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.607116] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356607, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.700070] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab17ee90-841e-4db4-acce-eafeffa97853 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.710110] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39429e8-c654-4ab4-aea7-de9bd9924b67 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.741024] env[61906]: DEBUG nova.network.neutron [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.743514] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6517a5cc-311e-4556-aeae-015a42698264 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.751464] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb2ac7b-fd47-4ce4-bdd8-45837ac26d77 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.766999] env[61906]: DEBUG nova.compute.provider_tree [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.770694] env[61906]: DEBUG oslo_concurrency.lockutils [None req-14157e94-0561-4df0-abbc-8f668db85662 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.332s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.830823] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.831092] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.831835] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.832084] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.832294] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.834852] env[61906]: INFO nova.compute.manager [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Terminating instance [ 821.837550] env[61906]: DEBUG nova.compute.manager [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 821.837653] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 821.837867] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5417552c-2eb8-44fc-a6cc-6e37b0454a96 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.845710] env[61906]: DEBUG oslo_vmware.api [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 821.845710] env[61906]: value = "task-1356608" [ 821.845710] env[61906]: _type = "Task" [ 821.845710] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.854745] env[61906]: DEBUG oslo_vmware.api [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356608, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.926390] env[61906]: DEBUG nova.network.neutron [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance_info_cache with network_info: [{"id": "d803c8e8-7145-412c-aa3e-2545666742eb", "address": "fa:16:3e:46:59:77", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd803c8e8-71", "ovs_interfaceid": "d803c8e8-7145-412c-aa3e-2545666742eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.107443] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356607, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448601} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.107740] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 7bae658e-b050-4639-b34e-c2671ef5c773/7bae658e-b050-4639-b34e-c2671ef5c773.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 822.107922] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 822.108584] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cd514d5b-dbd5-4fc3-a357-db52243c1be9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.114688] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 822.114688] env[61906]: value = "task-1356609" [ 822.114688] env[61906]: _type = "Task" [ 822.114688] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.124127] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356609, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.273095] env[61906]: DEBUG nova.scheduler.client.report [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.355531] env[61906]: DEBUG oslo_vmware.api [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356608, 'name': PowerOffVM_Task, 'duration_secs': 0.281004} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.355806] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 822.356022] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 822.356224] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288960', 'volume_id': '96c5a40a-fe62-4fbf-8378-5fa3c290af1a', 'name': 'volume-96c5a40a-fe62-4fbf-8378-5fa3c290af1a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9a23071d-40fd-4446-aa03-ecbddcafe3d6', 'attached_at': '', 'detached_at': '', 'volume_id': '96c5a40a-fe62-4fbf-8378-5fa3c290af1a', 'serial': '96c5a40a-fe62-4fbf-8378-5fa3c290af1a'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 822.357019] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa0a25fb-ee6e-41cc-8d3d-69cf5c83b153 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.377920] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31357361-9e7e-422e-a41c-e820747e499f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.384440] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82917ea-c7ee-4986-9006-5254ed74011f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.404565] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2ea825-06d8-4c81-a979-8b97fca24a8d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.419730] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] The volume has not been displaced from its original location: [datastore2] volume-96c5a40a-fe62-4fbf-8378-5fa3c290af1a/volume-96c5a40a-fe62-4fbf-8378-5fa3c290af1a.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 822.424829] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Reconfiguring VM instance instance-00000041 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 822.425099] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41614d63-aa6d-45e1-987d-85dc868aadc1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.437538] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.437788] env[61906]: DEBUG nova.compute.manager [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Instance network_info: |[{"id": "d803c8e8-7145-412c-aa3e-2545666742eb", "address": "fa:16:3e:46:59:77", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd803c8e8-71", "ovs_interfaceid": "d803c8e8-7145-412c-aa3e-2545666742eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 822.438162] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 822.438511] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:59:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8233b50c-be21-441a-a357-a29138a95b8b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd803c8e8-7145-412c-aa3e-2545666742eb', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.445507] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Creating folder: Project (0c6414fa082f49c2a568a772a256a8e5). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.445726] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 822.447252] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-752b781a-9b1d-4016-b5c7-cb4cb48f2eb3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.453040] env[61906]: DEBUG oslo_vmware.api [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 822.453040] env[61906]: value = "task-1356610" [ 822.453040] env[61906]: _type = "Task" [ 822.453040] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.458138] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Created folder: Project (0c6414fa082f49c2a568a772a256a8e5) in parent group-v288914. [ 822.458272] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Creating folder: Instances. Parent ref: group-v288965. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.459099] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c65ec3f7-06c3-446a-99a2-cd13c8112483 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.465081] env[61906]: DEBUG oslo_vmware.api [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356610, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.474023] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Created folder: Instances in parent group-v288965. [ 822.474115] env[61906]: DEBUG oslo.service.loopingcall [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.474649] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e133605d-c630-4b9e-a314-bf496c853710] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 822.474978] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3787c66f-37c6-43b7-8bb2-7344e14a4942 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.495905] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.495905] env[61906]: value = "task-1356613" [ 822.495905] env[61906]: _type = "Task" [ 822.495905] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.502051] env[61906]: DEBUG nova.compute.manager [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 822.506983] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356613, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.528557] env[61906]: DEBUG nova.virt.hardware [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 822.528817] env[61906]: DEBUG nova.virt.hardware [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 822.528992] env[61906]: DEBUG nova.virt.hardware [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 822.529210] env[61906]: DEBUG nova.virt.hardware [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 822.529360] env[61906]: DEBUG nova.virt.hardware [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 822.529526] env[61906]: DEBUG nova.virt.hardware [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 822.530554] env[61906]: DEBUG nova.virt.hardware [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 822.530554] env[61906]: DEBUG nova.virt.hardware [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 822.530554] env[61906]: DEBUG nova.virt.hardware [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 822.530554] env[61906]: DEBUG nova.virt.hardware [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 822.530554] env[61906]: DEBUG nova.virt.hardware [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 822.531407] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c717cf1-9439-4568-9c72-d138bdad5c20 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.539137] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8ea801-cd54-4461-9161-0fb7ebfeb963 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.623327] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356609, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06007} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.623606] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 822.624390] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157aa636-d9d1-46eb-a5f2-58fcbc198e90 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.646485] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 7bae658e-b050-4639-b34e-c2671ef5c773/7bae658e-b050-4639-b34e-c2671ef5c773.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.647787] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc0fa0af-31fa-4116-9255-15e579804ec2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.662588] env[61906]: DEBUG nova.compute.manager [req-b2164909-9979-4922-a8d0-e8729908ed62 req-05b72055-97e2-4b97-a175-2adced13d3ae service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Received event network-changed-d803c8e8-7145-412c-aa3e-2545666742eb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.662766] env[61906]: DEBUG nova.compute.manager [req-b2164909-9979-4922-a8d0-e8729908ed62 req-05b72055-97e2-4b97-a175-2adced13d3ae service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Refreshing instance network info cache due to event network-changed-d803c8e8-7145-412c-aa3e-2545666742eb. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 822.662975] env[61906]: DEBUG oslo_concurrency.lockutils [req-b2164909-9979-4922-a8d0-e8729908ed62 req-05b72055-97e2-4b97-a175-2adced13d3ae service nova] Acquiring lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.663130] env[61906]: DEBUG oslo_concurrency.lockutils [req-b2164909-9979-4922-a8d0-e8729908ed62 req-05b72055-97e2-4b97-a175-2adced13d3ae service nova] Acquired lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.663297] env[61906]: DEBUG nova.network.neutron [req-b2164909-9979-4922-a8d0-e8729908ed62 req-05b72055-97e2-4b97-a175-2adced13d3ae service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Refreshing network info cache for port d803c8e8-7145-412c-aa3e-2545666742eb {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 822.669420] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 822.669420] env[61906]: value = "task-1356614" [ 822.669420] env[61906]: _type = "Task" [ 822.669420] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.678456] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356614, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.778282] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.778812] env[61906]: DEBUG nova.compute.manager [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 822.781947] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.915s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.782214] env[61906]: DEBUG nova.objects.instance [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lazy-loading 'resources' on Instance uuid 7943ab1a-82df-48dd-874c-15dc015db51e {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 822.931380] env[61906]: DEBUG nova.network.neutron [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Successfully updated port: 96711ac0-bf70-468e-9ecc-70a5313e4bee {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.955112] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 822.955112] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 822.955112] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 822.964803] env[61906]: DEBUG oslo_vmware.api [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356610, 'name': ReconfigVM_Task, 'duration_secs': 0.209445} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.965209] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Reconfigured VM instance instance-00000041 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 822.969976] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b809798c-a186-4286-b82d-84a1d65ce24f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.985764] env[61906]: DEBUG oslo_vmware.api [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 822.985764] env[61906]: value = "task-1356615" [ 822.985764] env[61906]: _type = "Task" [ 822.985764] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.997842] env[61906]: DEBUG oslo_vmware.api [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356615, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.005769] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356613, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.180398] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356614, 'name': ReconfigVM_Task, 'duration_secs': 0.346098} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.180717] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 7bae658e-b050-4639-b34e-c2671ef5c773/7bae658e-b050-4639-b34e-c2671ef5c773.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.181354] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f962dea-82b0-44e0-a31f-f490f56a4cf5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.187780] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 823.187780] env[61906]: value = "task-1356616" [ 823.187780] env[61906]: _type = "Task" [ 823.187780] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.195700] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356616, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.285417] env[61906]: DEBUG nova.compute.utils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 823.289504] env[61906]: DEBUG nova.compute.manager [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 823.289673] env[61906]: DEBUG nova.network.neutron [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 823.364328] env[61906]: DEBUG nova.policy [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ec22db431334e7886acb497ebfc7eb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ac9a6686e7184fb1913e84a77985b449', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 823.420568] env[61906]: DEBUG nova.network.neutron [req-b2164909-9979-4922-a8d0-e8729908ed62 req-05b72055-97e2-4b97-a175-2adced13d3ae service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updated VIF entry in instance network info cache for port d803c8e8-7145-412c-aa3e-2545666742eb. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 823.420940] env[61906]: DEBUG nova.network.neutron [req-b2164909-9979-4922-a8d0-e8729908ed62 req-05b72055-97e2-4b97-a175-2adced13d3ae service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance_info_cache with network_info: [{"id": "d803c8e8-7145-412c-aa3e-2545666742eb", "address": "fa:16:3e:46:59:77", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd803c8e8-71", "ovs_interfaceid": "d803c8e8-7145-412c-aa3e-2545666742eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.433719] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Acquiring lock "refresh_cache-0552e012-edcf-444c-a9b3-f1cea93b9e82" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.433868] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Acquired lock "refresh_cache-0552e012-edcf-444c-a9b3-f1cea93b9e82" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.434031] env[61906]: DEBUG nova.network.neutron [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.464608] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Skipping network cache update for instance because it is being deleted. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 823.464786] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 823.464915] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: e133605d-c630-4b9e-a314-bf496c853710] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 823.465049] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 823.465175] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 823.481385] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.482785] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquired lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.482785] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 823.482785] env[61906]: DEBUG nova.objects.instance [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lazy-loading 'info_cache' on Instance uuid 7943ab1a-82df-48dd-874c-15dc015db51e {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 823.499121] env[61906]: DEBUG oslo_vmware.api [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356615, 'name': ReconfigVM_Task, 'duration_secs': 0.276151} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.502117] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288960', 'volume_id': '96c5a40a-fe62-4fbf-8378-5fa3c290af1a', 'name': 'volume-96c5a40a-fe62-4fbf-8378-5fa3c290af1a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9a23071d-40fd-4446-aa03-ecbddcafe3d6', 'attached_at': '', 'detached_at': '', 'volume_id': '96c5a40a-fe62-4fbf-8378-5fa3c290af1a', 'serial': '96c5a40a-fe62-4fbf-8378-5fa3c290af1a'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 823.502416] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 823.505242] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d8a8f7-d176-47da-94dd-3cea0864d1f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.513598] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356613, 'name': CreateVM_Task, 'duration_secs': 0.595889} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.515573] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e133605d-c630-4b9e-a314-bf496c853710] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 823.515868] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 823.516562] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.516925] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.517036] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 823.517571] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7692ebd4-a64d-438e-93d2-bf02000f3ccb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.518995] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57181009-ee29-4c4d-b782-3f8bc815e0ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.522028] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3428540-3308-4a3d-aebb-f40382722c58 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.527959] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 823.527959] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5213005c-9b4e-ee54-df78-24989af62503" [ 823.527959] env[61906]: _type = "Task" [ 823.527959] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.537850] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5213005c-9b4e-ee54-df78-24989af62503, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.540404] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77eac3b2-1d45-4c81-9359-93469da0eb47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.571970] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce4aaf0-0c05-4a52-8543-1f5f53335957 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.579619] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd717bb-8583-457b-bd37-02134aceba9e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.596353] env[61906]: DEBUG nova.compute.provider_tree [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.615573] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 823.615796] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 823.615976] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleting the datastore file [datastore2] 9a23071d-40fd-4446-aa03-ecbddcafe3d6 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 823.616283] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db1d2579-7e25-46db-872f-3235db958dcb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.623116] env[61906]: DEBUG oslo_vmware.api [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 823.623116] env[61906]: value = "task-1356618" [ 823.623116] env[61906]: _type = "Task" [ 823.623116] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.631796] env[61906]: DEBUG oslo_vmware.api [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356618, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.668646] env[61906]: DEBUG nova.network.neutron [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Successfully created port: b670d2ba-324d-4d95-b044-de1cc7c68bd5 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.697856] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356616, 'name': Rename_Task, 'duration_secs': 0.171122} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.698497] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 823.698744] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0fd15b5-4db5-431e-89e6-0fc1b176d28c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.704836] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 823.704836] env[61906]: value = "task-1356619" [ 823.704836] env[61906]: _type = "Task" [ 823.704836] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.712928] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356619, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.794335] env[61906]: DEBUG nova.compute.manager [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 823.926689] env[61906]: DEBUG oslo_concurrency.lockutils [req-b2164909-9979-4922-a8d0-e8729908ed62 req-05b72055-97e2-4b97-a175-2adced13d3ae service nova] Releasing lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.967271] env[61906]: DEBUG nova.network.neutron [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.040234] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5213005c-9b4e-ee54-df78-24989af62503, 'name': SearchDatastore_Task, 'duration_secs': 0.012381} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.040625] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.040966] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.041253] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.041451] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.041691] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.041995] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4d48f60-601d-4621-be02-b4d7a91f465f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.050908] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 824.051102] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 824.051807] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52850b20-0d2a-42a4-b30f-311a84424e19 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.056843] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 824.056843] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52e8215e-ff1f-64a3-9001-dcc179a64b1f" [ 824.056843] env[61906]: _type = "Task" [ 824.056843] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.064662] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52e8215e-ff1f-64a3-9001-dcc179a64b1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.102396] env[61906]: DEBUG nova.scheduler.client.report [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 824.132509] env[61906]: DEBUG oslo_vmware.api [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356618, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198538} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.132793] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 824.132978] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 824.133173] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 824.133351] env[61906]: INFO nova.compute.manager [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Took 2.30 seconds to destroy the instance on the hypervisor. [ 824.133589] env[61906]: DEBUG oslo.service.loopingcall [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.133778] env[61906]: DEBUG nova.compute.manager [-] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 824.133947] env[61906]: DEBUG nova.network.neutron [-] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 824.177874] env[61906]: DEBUG nova.network.neutron [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Updating instance_info_cache with network_info: [{"id": "96711ac0-bf70-468e-9ecc-70a5313e4bee", "address": "fa:16:3e:64:a0:62", "network": {"id": "65476ec7-1ddc-4b0f-86bc-090782193282", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-990976432-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ec67c1b35414992a70b6dfa372ac12e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496faa4d-d874-449b-905e-328ddd60b31b", "external-id": "nsx-vlan-transportzone-391", "segmentation_id": 391, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96711ac0-bf", "ovs_interfaceid": "96711ac0-bf70-468e-9ecc-70a5313e4bee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.218833] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356619, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.517746] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.567024] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52e8215e-ff1f-64a3-9001-dcc179a64b1f, 'name': SearchDatastore_Task, 'duration_secs': 0.008399} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.567825] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43c48d6d-f10a-49cd-8515-6affc6139937 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.572805] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 824.572805] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f29bbc-8436-1b4a-1432-74eb02e5ab83" [ 824.572805] env[61906]: _type = "Task" [ 824.572805] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.580740] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f29bbc-8436-1b4a-1432-74eb02e5ab83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.605205] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.824s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.607748] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.533s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.609397] env[61906]: INFO nova.compute.claims [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.632615] env[61906]: INFO nova.scheduler.client.report [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Deleted allocations for instance 7943ab1a-82df-48dd-874c-15dc015db51e [ 824.670665] env[61906]: DEBUG nova.compute.manager [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Received event network-vif-plugged-96711ac0-bf70-468e-9ecc-70a5313e4bee {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.670824] env[61906]: DEBUG oslo_concurrency.lockutils [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] Acquiring lock "0552e012-edcf-444c-a9b3-f1cea93b9e82-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.671046] env[61906]: DEBUG oslo_concurrency.lockutils [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] Lock "0552e012-edcf-444c-a9b3-f1cea93b9e82-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.671218] env[61906]: DEBUG oslo_concurrency.lockutils [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] Lock "0552e012-edcf-444c-a9b3-f1cea93b9e82-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.671399] env[61906]: DEBUG nova.compute.manager [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] No waiting events found dispatching network-vif-plugged-96711ac0-bf70-468e-9ecc-70a5313e4bee {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 824.671596] env[61906]: WARNING nova.compute.manager [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Received unexpected event network-vif-plugged-96711ac0-bf70-468e-9ecc-70a5313e4bee for instance with vm_state building and task_state spawning. [ 824.671757] env[61906]: DEBUG nova.compute.manager [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Received event network-changed-96711ac0-bf70-468e-9ecc-70a5313e4bee {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.671908] env[61906]: DEBUG nova.compute.manager [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Refreshing instance network info cache due to event network-changed-96711ac0-bf70-468e-9ecc-70a5313e4bee. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 824.672083] env[61906]: DEBUG oslo_concurrency.lockutils [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] Acquiring lock "refresh_cache-0552e012-edcf-444c-a9b3-f1cea93b9e82" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.680759] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Releasing lock "refresh_cache-0552e012-edcf-444c-a9b3-f1cea93b9e82" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.681135] env[61906]: DEBUG nova.compute.manager [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Instance network_info: |[{"id": "96711ac0-bf70-468e-9ecc-70a5313e4bee", "address": "fa:16:3e:64:a0:62", "network": {"id": "65476ec7-1ddc-4b0f-86bc-090782193282", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-990976432-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ec67c1b35414992a70b6dfa372ac12e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496faa4d-d874-449b-905e-328ddd60b31b", "external-id": "nsx-vlan-transportzone-391", "segmentation_id": 391, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96711ac0-bf", "ovs_interfaceid": "96711ac0-bf70-468e-9ecc-70a5313e4bee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 824.681591] env[61906]: DEBUG oslo_concurrency.lockutils [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] Acquired lock "refresh_cache-0552e012-edcf-444c-a9b3-f1cea93b9e82" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.681768] env[61906]: DEBUG nova.network.neutron [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Refreshing network info cache for port 96711ac0-bf70-468e-9ecc-70a5313e4bee {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 824.683083] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:a0:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '496faa4d-d874-449b-905e-328ddd60b31b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '96711ac0-bf70-468e-9ecc-70a5313e4bee', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 824.690794] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Creating folder: Project (2ec67c1b35414992a70b6dfa372ac12e). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.694319] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d42f1596-cccc-43c9-a736-4d7dc1129898 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.705022] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Created folder: Project (2ec67c1b35414992a70b6dfa372ac12e) in parent group-v288914. [ 824.705459] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Creating folder: Instances. Parent ref: group-v288968. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.705459] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b2ec3d40-593b-4b55-9544-3025af759830 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.715454] env[61906]: DEBUG oslo_vmware.api [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356619, 'name': PowerOnVM_Task, 'duration_secs': 0.542515} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.715715] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 824.715910] env[61906]: INFO nova.compute.manager [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Took 8.88 seconds to spawn the instance on the hypervisor. [ 824.716147] env[61906]: DEBUG nova.compute.manager [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 824.716880] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c0233e-a195-46cf-ba90-2a9a4ac52795 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.720207] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Created folder: Instances in parent group-v288968. [ 824.720406] env[61906]: DEBUG oslo.service.loopingcall [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.720892] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 824.721099] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-39be4991-20d9-496e-8cb1-21b63a3c1b3e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.746258] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 824.746258] env[61906]: value = "task-1356622" [ 824.746258] env[61906]: _type = "Task" [ 824.746258] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.754254] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356622, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.803068] env[61906]: DEBUG nova.compute.manager [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 824.831166] env[61906]: DEBUG nova.virt.hardware [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.831454] env[61906]: DEBUG nova.virt.hardware [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.831628] env[61906]: DEBUG nova.virt.hardware [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.831817] env[61906]: DEBUG nova.virt.hardware [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.831962] env[61906]: DEBUG nova.virt.hardware [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.832124] env[61906]: DEBUG nova.virt.hardware [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.832363] env[61906]: DEBUG nova.virt.hardware [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.832539] env[61906]: DEBUG nova.virt.hardware [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.832711] env[61906]: DEBUG nova.virt.hardware [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.832875] env[61906]: DEBUG nova.virt.hardware [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.833082] env[61906]: DEBUG nova.virt.hardware [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.833982] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a2218d-e60a-4436-9fd5-42b13eb01c8e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.844312] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33d57e3-53f4-47f1-9ea8-d5470d3fe8d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.951032] env[61906]: DEBUG nova.network.neutron [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Updated VIF entry in instance network info cache for port 96711ac0-bf70-468e-9ecc-70a5313e4bee. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 824.951419] env[61906]: DEBUG nova.network.neutron [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Updating instance_info_cache with network_info: [{"id": "96711ac0-bf70-468e-9ecc-70a5313e4bee", "address": "fa:16:3e:64:a0:62", "network": {"id": "65476ec7-1ddc-4b0f-86bc-090782193282", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-990976432-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ec67c1b35414992a70b6dfa372ac12e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496faa4d-d874-449b-905e-328ddd60b31b", "external-id": "nsx-vlan-transportzone-391", "segmentation_id": 391, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96711ac0-bf", "ovs_interfaceid": "96711ac0-bf70-468e-9ecc-70a5313e4bee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.010854] env[61906]: DEBUG nova.network.neutron [-] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.083778] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f29bbc-8436-1b4a-1432-74eb02e5ab83, 'name': SearchDatastore_Task, 'duration_secs': 0.017876} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.084212] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.084594] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] e133605d-c630-4b9e-a314-bf496c853710/e133605d-c630-4b9e-a314-bf496c853710.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 825.085012] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f0ac735-ee11-41a8-9740-0502c5df4231 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.094021] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 825.094021] env[61906]: value = "task-1356623" [ 825.094021] env[61906]: _type = "Task" [ 825.094021] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.099819] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356623, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.105850] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.140983] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1c48945e-1f99-433a-b825-2637f44f169e tempest-SecurityGroupsTestJSON-1679810606 tempest-SecurityGroupsTestJSON-1679810606-project-member] Lock "7943ab1a-82df-48dd-874c-15dc015db51e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.267s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.146910] env[61906]: DEBUG nova.compute.manager [req-58e1a23e-baac-4a18-8055-6420cd6d343f req-cd17787c-80ff-42a2-9d24-e197977950c4 service nova] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Received event network-vif-plugged-b670d2ba-324d-4d95-b044-de1cc7c68bd5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.147242] env[61906]: DEBUG oslo_concurrency.lockutils [req-58e1a23e-baac-4a18-8055-6420cd6d343f req-cd17787c-80ff-42a2-9d24-e197977950c4 service nova] Acquiring lock "24b4a747-67aa-4388-aed6-cb646cd55765-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.148147] env[61906]: DEBUG oslo_concurrency.lockutils [req-58e1a23e-baac-4a18-8055-6420cd6d343f req-cd17787c-80ff-42a2-9d24-e197977950c4 service nova] Lock "24b4a747-67aa-4388-aed6-cb646cd55765-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.148396] env[61906]: DEBUG oslo_concurrency.lockutils [req-58e1a23e-baac-4a18-8055-6420cd6d343f req-cd17787c-80ff-42a2-9d24-e197977950c4 service nova] Lock "24b4a747-67aa-4388-aed6-cb646cd55765-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.148621] env[61906]: DEBUG nova.compute.manager [req-58e1a23e-baac-4a18-8055-6420cd6d343f req-cd17787c-80ff-42a2-9d24-e197977950c4 service nova] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] No waiting events found dispatching network-vif-plugged-b670d2ba-324d-4d95-b044-de1cc7c68bd5 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 825.148843] env[61906]: WARNING nova.compute.manager [req-58e1a23e-baac-4a18-8055-6420cd6d343f req-cd17787c-80ff-42a2-9d24-e197977950c4 service nova] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Received unexpected event network-vif-plugged-b670d2ba-324d-4d95-b044-de1cc7c68bd5 for instance with vm_state building and task_state spawning. [ 825.253176] env[61906]: INFO nova.compute.manager [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Took 28.66 seconds to build instance. [ 825.258371] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356622, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.309629] env[61906]: DEBUG nova.network.neutron [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Successfully updated port: b670d2ba-324d-4d95-b044-de1cc7c68bd5 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.454369] env[61906]: DEBUG oslo_concurrency.lockutils [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] Releasing lock "refresh_cache-0552e012-edcf-444c-a9b3-f1cea93b9e82" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.454639] env[61906]: DEBUG nova.compute.manager [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Received event network-vif-deleted-d962bee4-e780-4b00-b521-255e4c0e7fad {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.454820] env[61906]: INFO nova.compute.manager [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Neutron deleted interface d962bee4-e780-4b00-b521-255e4c0e7fad; detaching it from the instance and deleting it from the info cache [ 825.454990] env[61906]: DEBUG nova.network.neutron [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.513556] env[61906]: INFO nova.compute.manager [-] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Took 1.38 seconds to deallocate network for instance. [ 825.602630] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356623, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.611027] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Releasing lock "refresh_cache-7943ab1a-82df-48dd-874c-15dc015db51e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.612041] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 825.612041] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.612041] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.612041] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.612041] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.612041] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.612357] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.612357] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 825.612497] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.660938] env[61906]: INFO nova.compute.manager [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Rescuing [ 825.661214] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "refresh_cache-7bae658e-b050-4639-b34e-c2671ef5c773" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.662128] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "refresh_cache-7bae658e-b050-4639-b34e-c2671ef5c773" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.662343] env[61906]: DEBUG nova.network.neutron [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 825.758357] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356622, 'name': CreateVM_Task, 'duration_secs': 0.529814} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.758859] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e2acd569-2eb5-488b-82b7-7f3eae05e149 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7bae658e-b050-4639-b34e-c2671ef5c773" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.692s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.759076] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 825.759775] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.759974] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.760383] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.760680] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc29c97c-a86e-416f-b2c2-eaaae5f0e5ca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.767178] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Waiting for the task: (returnval){ [ 825.767178] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528115d5-a3b1-fc50-5069-7985c830c4fe" [ 825.767178] env[61906]: _type = "Task" [ 825.767178] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.776027] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528115d5-a3b1-fc50-5069-7985c830c4fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.808507] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b0068d-7b32-4f23-abeb-17267dde1cde {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.811356] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "refresh_cache-24b4a747-67aa-4388-aed6-cb646cd55765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.811492] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "refresh_cache-24b4a747-67aa-4388-aed6-cb646cd55765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.811641] env[61906]: DEBUG nova.network.neutron [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 825.818398] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7020f6-e5ba-4bb1-8647-10c84f15c340 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.850671] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e636a7b7-3a1a-403a-9e1b-6744157b2475 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.858578] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5359f690-32df-4f9e-b217-8e9c216e89d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.872617] env[61906]: DEBUG nova.compute.provider_tree [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.958038] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a2fe93e2-c326-4d3d-b749-5f484940fa01 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.967276] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0fd1db1-99b0-479d-a872-534f8881eb0d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.993330] env[61906]: DEBUG nova.compute.manager [req-279e6e90-7886-4d25-8558-d5cae6cca8b4 req-d6e1c56c-f085-4f60-a25e-3baf48b802b3 service nova] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Detach interface failed, port_id=d962bee4-e780-4b00-b521-255e4c0e7fad, reason: Instance 9a23071d-40fd-4446-aa03-ecbddcafe3d6 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 826.072705] env[61906]: INFO nova.compute.manager [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Took 0.56 seconds to detach 1 volumes for instance. [ 826.107598] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356623, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516526} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.108030] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] e133605d-c630-4b9e-a314-bf496c853710/e133605d-c630-4b9e-a314-bf496c853710.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 826.108291] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 826.108662] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-87961322-3cfe-44a9-a0b8-651768479b5b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.116375] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.118767] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 826.118767] env[61906]: value = "task-1356624" [ 826.118767] env[61906]: _type = "Task" [ 826.118767] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.131129] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356624, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.277868] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528115d5-a3b1-fc50-5069-7985c830c4fe, 'name': SearchDatastore_Task, 'duration_secs': 0.051065} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.280014] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.280257] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.280492] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.280638] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.280813] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.281087] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ef665a8-9cb3-4e25-af3f-e22b6145ba03 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.293717] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.293898] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 826.294624] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12e016f4-6fbe-47c9-ac65-19264981e65d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.299592] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Waiting for the task: (returnval){ [ 826.299592] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5270b5fc-dc50-095c-53ad-9ba199837eef" [ 826.299592] env[61906]: _type = "Task" [ 826.299592] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.308692] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5270b5fc-dc50-095c-53ad-9ba199837eef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.375829] env[61906]: DEBUG nova.scheduler.client.report [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 826.398899] env[61906]: DEBUG nova.network.neutron [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.449913] env[61906]: DEBUG nova.network.neutron [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Updating instance_info_cache with network_info: [{"id": "4c806e0c-9874-4943-9a71-05fc2c40045e", "address": "fa:16:3e:55:c4:86", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c806e0c-98", "ovs_interfaceid": "4c806e0c-9874-4943-9a71-05fc2c40045e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.579718] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.629692] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356624, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067194} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.629971] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.630774] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b826b74-032a-4d90-843f-7f12f3dc997f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.652855] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] e133605d-c630-4b9e-a314-bf496c853710/e133605d-c630-4b9e-a314-bf496c853710.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.653155] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b97ac9e4-8168-4a83-ad6d-1dd6b6f32f20 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.674600] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 826.674600] env[61906]: value = "task-1356625" [ 826.674600] env[61906]: _type = "Task" [ 826.674600] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.683011] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356625, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.809898] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5270b5fc-dc50-095c-53ad-9ba199837eef, 'name': SearchDatastore_Task, 'duration_secs': 0.029031} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.810730] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35b74627-9e42-47aa-87c2-e1b11622da8c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.815943] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Waiting for the task: (returnval){ [ 826.815943] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52dcef02-c470-9e7d-17a8-a78df38f8842" [ 826.815943] env[61906]: _type = "Task" [ 826.815943] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.827107] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52dcef02-c470-9e7d-17a8-a78df38f8842, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.880506] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.273s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.881041] env[61906]: DEBUG nova.compute.manager [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.883960] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.362s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.885596] env[61906]: INFO nova.compute.claims [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.891284] env[61906]: DEBUG nova.network.neutron [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Updating instance_info_cache with network_info: [{"id": "b670d2ba-324d-4d95-b044-de1cc7c68bd5", "address": "fa:16:3e:31:95:05", "network": {"id": "5b29c9e0-23c9-42ee-9be4-60178bddedbe", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643906254-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac9a6686e7184fb1913e84a77985b449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb670d2ba-32", "ovs_interfaceid": "b670d2ba-324d-4d95-b044-de1cc7c68bd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.953217] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "refresh_cache-7bae658e-b050-4639-b34e-c2671ef5c773" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.174902] env[61906]: DEBUG nova.compute.manager [req-447bfdc9-93e6-46d7-bc42-27b3fa9c9a9b req-2c485037-ab2d-4e70-913d-6c035376522c service nova] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Received event network-changed-b670d2ba-324d-4d95-b044-de1cc7c68bd5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.175150] env[61906]: DEBUG nova.compute.manager [req-447bfdc9-93e6-46d7-bc42-27b3fa9c9a9b req-2c485037-ab2d-4e70-913d-6c035376522c service nova] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Refreshing instance network info cache due to event network-changed-b670d2ba-324d-4d95-b044-de1cc7c68bd5. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 827.175352] env[61906]: DEBUG oslo_concurrency.lockutils [req-447bfdc9-93e6-46d7-bc42-27b3fa9c9a9b req-2c485037-ab2d-4e70-913d-6c035376522c service nova] Acquiring lock "refresh_cache-24b4a747-67aa-4388-aed6-cb646cd55765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.185945] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356625, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.327977] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52dcef02-c470-9e7d-17a8-a78df38f8842, 'name': SearchDatastore_Task, 'duration_secs': 0.013252} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.328321] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.328521] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 0552e012-edcf-444c-a9b3-f1cea93b9e82/0552e012-edcf-444c-a9b3-f1cea93b9e82.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 827.328824] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ef5b582-b13c-43b2-9c1a-45beac22becd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.336816] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Waiting for the task: (returnval){ [ 827.336816] env[61906]: value = "task-1356626" [ 827.336816] env[61906]: _type = "Task" [ 827.336816] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.344940] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356626, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.390592] env[61906]: DEBUG nova.compute.utils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.393042] env[61906]: DEBUG nova.compute.manager [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 827.393237] env[61906]: DEBUG nova.network.neutron [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 827.397495] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "refresh_cache-24b4a747-67aa-4388-aed6-cb646cd55765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.397761] env[61906]: DEBUG nova.compute.manager [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Instance network_info: |[{"id": "b670d2ba-324d-4d95-b044-de1cc7c68bd5", "address": "fa:16:3e:31:95:05", "network": {"id": "5b29c9e0-23c9-42ee-9be4-60178bddedbe", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643906254-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac9a6686e7184fb1913e84a77985b449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb670d2ba-32", "ovs_interfaceid": "b670d2ba-324d-4d95-b044-de1cc7c68bd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 827.398041] env[61906]: DEBUG oslo_concurrency.lockutils [req-447bfdc9-93e6-46d7-bc42-27b3fa9c9a9b req-2c485037-ab2d-4e70-913d-6c035376522c service nova] Acquired lock "refresh_cache-24b4a747-67aa-4388-aed6-cb646cd55765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.398218] env[61906]: DEBUG nova.network.neutron [req-447bfdc9-93e6-46d7-bc42-27b3fa9c9a9b req-2c485037-ab2d-4e70-913d-6c035376522c service nova] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Refreshing network info cache for port b670d2ba-324d-4d95-b044-de1cc7c68bd5 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 827.399246] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:95:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c02dd284-ab80-451c-93eb-48c8360acb9c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b670d2ba-324d-4d95-b044-de1cc7c68bd5', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 827.408868] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Creating folder: Project (ac9a6686e7184fb1913e84a77985b449). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.410725] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-11e52672-5718-4a4c-8133-7473eb23b34e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.421868] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Created folder: Project (ac9a6686e7184fb1913e84a77985b449) in parent group-v288914. [ 827.422076] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Creating folder: Instances. Parent ref: group-v288971. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.422313] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f70f27be-5cd5-4970-aa5d-b1381175879a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.431349] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Created folder: Instances in parent group-v288971. [ 827.432126] env[61906]: DEBUG oslo.service.loopingcall [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.432126] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 827.432126] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82f98b4b-3aca-40d1-b317-c23d99ea2195 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.452596] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 827.452596] env[61906]: value = "task-1356629" [ 827.452596] env[61906]: _type = "Task" [ 827.452596] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.460748] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356629, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.474118] env[61906]: DEBUG nova.policy [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f084b58b30a64e06ad69e8f6b366207c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f3bb9ce83b3430bb9202d3c72e77b4c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 827.485839] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 827.486182] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e8460ed-a7d2-4165-8251-62acb5adfdc8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.492290] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 827.492290] env[61906]: value = "task-1356630" [ 827.492290] env[61906]: _type = "Task" [ 827.492290] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.501958] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356630, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.690158] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356625, 'name': ReconfigVM_Task, 'duration_secs': 0.684365} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.691132] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Reconfigured VM instance instance-00000046 to attach disk [datastore1] e133605d-c630-4b9e-a314-bf496c853710/e133605d-c630-4b9e-a314-bf496c853710.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.691892] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f743c053-20e2-4131-abd3-247a72cac554 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.699106] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 827.699106] env[61906]: value = "task-1356631" [ 827.699106] env[61906]: _type = "Task" [ 827.699106] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.708457] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356631, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.848310] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356626, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.894804] env[61906]: DEBUG nova.compute.manager [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.965737] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356629, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.007372] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356630, 'name': PowerOffVM_Task, 'duration_secs': 0.235686} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.007649] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 828.008503] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4afbcbf-b986-4d91-a8c7-d113dd1bbec3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.043984] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98fb6a4-f4f1-4593-bcfa-471aa2b053ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.093948] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 828.094658] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-409d0471-c682-48ab-82d1-4b58208655d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.103214] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 828.103214] env[61906]: value = "task-1356632" [ 828.103214] env[61906]: _type = "Task" [ 828.103214] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.116441] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] VM already powered off {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 828.116441] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.116441] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.116441] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.116441] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.117992] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-635899e9-2acf-4a54-bb2a-120525539886 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.126873] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.127621] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 828.127860] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0132441-ced5-4536-a040-5a92e1ecfd64 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.138549] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 828.138549] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52760337-2566-63ae-a3ae-591fec3747e2" [ 828.138549] env[61906]: _type = "Task" [ 828.138549] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.147959] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52760337-2566-63ae-a3ae-591fec3747e2, 'name': SearchDatastore_Task, 'duration_secs': 0.008919} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.149098] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca796f36-a499-4fa6-bdc5-de685c8283e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.154832] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 828.154832] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ff1aa1-7988-0ada-81a0-0e0fb4e4c12f" [ 828.154832] env[61906]: _type = "Task" [ 828.154832] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.170524] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ff1aa1-7988-0ada-81a0-0e0fb4e4c12f, 'name': SearchDatastore_Task, 'duration_secs': 0.009468} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.170793] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.171066] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 7bae658e-b050-4639-b34e-c2671ef5c773/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1-rescue.vmdk. {{(pid=61906) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 828.171365] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e2b7562-d58d-4fec-963b-d2a6ba5a14bc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.178729] env[61906]: DEBUG nova.network.neutron [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Successfully created port: f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 828.182735] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 828.182735] env[61906]: value = "task-1356633" [ 828.182735] env[61906]: _type = "Task" [ 828.182735] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.194288] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356633, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.208471] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356631, 'name': Rename_Task, 'duration_secs': 0.179702} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.208759] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 828.209027] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c98ebd2-3890-430d-8eba-a9c6737f8555 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.216420] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 828.216420] env[61906]: value = "task-1356634" [ 828.216420] env[61906]: _type = "Task" [ 828.216420] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.224628] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356634, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.227892] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a538bd29-0668-4f58-815a-767673450770 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.234879] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968ea7bf-b7a1-452b-b47f-9b7021315dfa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.267247] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a2ea0f-1811-4783-84d1-0f384b6d4b9b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.276216] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963e1261-88d0-4017-b140-302eee01a427 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.292570] env[61906]: DEBUG nova.compute.provider_tree [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.348936] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356626, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526826} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.349322] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 0552e012-edcf-444c-a9b3-f1cea93b9e82/0552e012-edcf-444c-a9b3-f1cea93b9e82.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.349503] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.349702] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ecc6842-c165-44ad-a9ad-0c33a8668729 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.356787] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Waiting for the task: (returnval){ [ 828.356787] env[61906]: value = "task-1356635" [ 828.356787] env[61906]: _type = "Task" [ 828.356787] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.372319] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356635, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.422180] env[61906]: DEBUG nova.network.neutron [req-447bfdc9-93e6-46d7-bc42-27b3fa9c9a9b req-2c485037-ab2d-4e70-913d-6c035376522c service nova] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Updated VIF entry in instance network info cache for port b670d2ba-324d-4d95-b044-de1cc7c68bd5. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 828.423200] env[61906]: DEBUG nova.network.neutron [req-447bfdc9-93e6-46d7-bc42-27b3fa9c9a9b req-2c485037-ab2d-4e70-913d-6c035376522c service nova] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Updating instance_info_cache with network_info: [{"id": "b670d2ba-324d-4d95-b044-de1cc7c68bd5", "address": "fa:16:3e:31:95:05", "network": {"id": "5b29c9e0-23c9-42ee-9be4-60178bddedbe", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643906254-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac9a6686e7184fb1913e84a77985b449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb670d2ba-32", "ovs_interfaceid": "b670d2ba-324d-4d95-b044-de1cc7c68bd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.467165] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356629, 'name': CreateVM_Task, 'duration_secs': 0.58853} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.467436] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 828.468350] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.468549] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.468931] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.469676] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59d74543-eb5b-4600-a4b2-d8f1d085ce9c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.475977] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 828.475977] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b0c159-27c4-f9f7-2dd1-6696246ece3c" [ 828.475977] env[61906]: _type = "Task" [ 828.475977] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.486480] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b0c159-27c4-f9f7-2dd1-6696246ece3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.692272] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356633, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439616} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.692556] env[61906]: INFO nova.virt.vmwareapi.ds_util [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 7bae658e-b050-4639-b34e-c2671ef5c773/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1-rescue.vmdk. [ 828.693334] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0319448-a1db-45c5-9509-2ed258893cf5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.717330] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 7bae658e-b050-4639-b34e-c2671ef5c773/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.717688] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c1acf03-0bd0-448b-a403-166444872bcb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.740036] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356634, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.740998] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 828.740998] env[61906]: value = "task-1356636" [ 828.740998] env[61906]: _type = "Task" [ 828.740998] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.748261] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356636, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.797419] env[61906]: DEBUG nova.scheduler.client.report [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.867159] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356635, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.27155} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.867475] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.868306] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4512f4-a6c9-45c5-86cc-ceff3b709ff8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.891417] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 0552e012-edcf-444c-a9b3-f1cea93b9e82/0552e012-edcf-444c-a9b3-f1cea93b9e82.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.891961] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ccb09fb-b004-4c96-a724-2a2b929b746c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.910900] env[61906]: DEBUG nova.compute.manager [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.914334] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Waiting for the task: (returnval){ [ 828.914334] env[61906]: value = "task-1356637" [ 828.914334] env[61906]: _type = "Task" [ 828.914334] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.923179] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356637, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.925812] env[61906]: DEBUG oslo_concurrency.lockutils [req-447bfdc9-93e6-46d7-bc42-27b3fa9c9a9b req-2c485037-ab2d-4e70-913d-6c035376522c service nova] Releasing lock "refresh_cache-24b4a747-67aa-4388-aed6-cb646cd55765" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.938315] env[61906]: DEBUG nova.virt.hardware [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.940526] env[61906]: DEBUG nova.virt.hardware [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.940526] env[61906]: DEBUG nova.virt.hardware [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.940526] env[61906]: DEBUG nova.virt.hardware [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.940526] env[61906]: DEBUG nova.virt.hardware [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.940526] env[61906]: DEBUG nova.virt.hardware [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.940526] env[61906]: DEBUG nova.virt.hardware [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.940526] env[61906]: DEBUG nova.virt.hardware [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.940526] env[61906]: DEBUG nova.virt.hardware [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.940526] env[61906]: DEBUG nova.virt.hardware [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.940526] env[61906]: DEBUG nova.virt.hardware [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.941218] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de53145-08bd-4328-9a8b-ce9a039816ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.949181] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9928d46c-903d-4a08-a968-b6fd84aaade3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.985977] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b0c159-27c4-f9f7-2dd1-6696246ece3c, 'name': SearchDatastore_Task, 'duration_secs': 0.019208} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.986339] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.986590] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.986847] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.986998] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.987213] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.987492] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0124c181-861f-4f00-9980-898ebd63568c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.996665] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.996760] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 828.997536] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-314a12b8-7120-4d5e-a575-d0d0a05fd76c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.003025] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 829.003025] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52edf104-ab10-3b22-8157-d390fe830828" [ 829.003025] env[61906]: _type = "Task" [ 829.003025] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.010654] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52edf104-ab10-3b22-8157-d390fe830828, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.228558] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356634, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.251569] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356636, 'name': ReconfigVM_Task, 'duration_secs': 0.318162} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.251921] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 7bae658e-b050-4639-b34e-c2671ef5c773/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.252775] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2861ecc-664b-49ed-9a08-adf93796cb16 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.278524] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7135597-fa25-49ae-9a8c-64c7f54ef4f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.293146] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 829.293146] env[61906]: value = "task-1356638" [ 829.293146] env[61906]: _type = "Task" [ 829.293146] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.303250] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.303781] env[61906]: DEBUG nova.compute.manager [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 829.306287] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356638, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.306563] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.239s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.306809] env[61906]: DEBUG nova.objects.instance [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lazy-loading 'resources' on Instance uuid 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 829.424883] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356637, 'name': ReconfigVM_Task, 'duration_secs': 0.335165} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.425883] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 0552e012-edcf-444c-a9b3-f1cea93b9e82/0552e012-edcf-444c-a9b3-f1cea93b9e82.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.428143] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0fad8730-9d0d-47cc-a4f4-10110ff1ebe0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.435547] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Waiting for the task: (returnval){ [ 829.435547] env[61906]: value = "task-1356639" [ 829.435547] env[61906]: _type = "Task" [ 829.435547] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.446784] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356639, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.514610] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52edf104-ab10-3b22-8157-d390fe830828, 'name': SearchDatastore_Task, 'duration_secs': 0.011867} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.515810] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62287c8b-a453-41bc-963b-a0bc90d6285c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.520905] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 829.520905] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ede75c-eb31-663f-f57d-66691566b107" [ 829.520905] env[61906]: _type = "Task" [ 829.520905] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.528647] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ede75c-eb31-663f-f57d-66691566b107, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.712297] env[61906]: DEBUG nova.compute.manager [req-daef38f3-33d5-4694-a754-5d6a439e1349 req-81d7aab4-3540-4013-a318-076138f3ca84 service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Received event network-vif-plugged-f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.712519] env[61906]: DEBUG oslo_concurrency.lockutils [req-daef38f3-33d5-4694-a754-5d6a439e1349 req-81d7aab4-3540-4013-a318-076138f3ca84 service nova] Acquiring lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.712728] env[61906]: DEBUG oslo_concurrency.lockutils [req-daef38f3-33d5-4694-a754-5d6a439e1349 req-81d7aab4-3540-4013-a318-076138f3ca84 service nova] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.712893] env[61906]: DEBUG oslo_concurrency.lockutils [req-daef38f3-33d5-4694-a754-5d6a439e1349 req-81d7aab4-3540-4013-a318-076138f3ca84 service nova] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.713068] env[61906]: DEBUG nova.compute.manager [req-daef38f3-33d5-4694-a754-5d6a439e1349 req-81d7aab4-3540-4013-a318-076138f3ca84 service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] No waiting events found dispatching network-vif-plugged-f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 829.713230] env[61906]: WARNING nova.compute.manager [req-daef38f3-33d5-4694-a754-5d6a439e1349 req-81d7aab4-3540-4013-a318-076138f3ca84 service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Received unexpected event network-vif-plugged-f119e802-0a46-4390-927d-805c2e23a4b4 for instance with vm_state building and task_state spawning. [ 829.728022] env[61906]: DEBUG oslo_vmware.api [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356634, 'name': PowerOnVM_Task, 'duration_secs': 1.076913} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.728022] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 829.728022] env[61906]: INFO nova.compute.manager [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Took 9.66 seconds to spawn the instance on the hypervisor. [ 829.728022] env[61906]: DEBUG nova.compute.manager [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.728794] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc129c53-e030-46a1-82ff-ad614aa80de7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.803988] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356638, 'name': ReconfigVM_Task, 'duration_secs': 0.154407} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.805025] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 829.806038] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-41c3f4cf-c5eb-40f2-8bec-c1ba7b21e40e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.809398] env[61906]: DEBUG nova.compute.utils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.812095] env[61906]: DEBUG nova.compute.manager [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 829.816825] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 829.816825] env[61906]: value = "task-1356640" [ 829.816825] env[61906]: _type = "Task" [ 829.816825] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.827130] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356640, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.829227] env[61906]: DEBUG nova.network.neutron [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Successfully updated port: f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.948507] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356639, 'name': Rename_Task, 'duration_secs': 0.483676} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.949076] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 829.949411] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f75b3fa1-76f4-4ce6-beeb-98a67ded613e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.957892] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Waiting for the task: (returnval){ [ 829.957892] env[61906]: value = "task-1356641" [ 829.957892] env[61906]: _type = "Task" [ 829.957892] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.970991] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356641, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.038742] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ede75c-eb31-663f-f57d-66691566b107, 'name': SearchDatastore_Task, 'duration_secs': 0.009734} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.039091] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.039366] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 24b4a747-67aa-4388-aed6-cb646cd55765/24b4a747-67aa-4388-aed6-cb646cd55765.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 830.039624] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b57db74a-7143-4ed8-9cc1-d345dd4171aa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.049893] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 830.049893] env[61906]: value = "task-1356642" [ 830.049893] env[61906]: _type = "Task" [ 830.049893] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.064097] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356642, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.084397] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c19c26-cdbc-46b0-af94-7b9b120dab08 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.092360] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f00cd63-b1f5-47e5-9783-086487a86b58 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.126517] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6f1f20-d32d-452c-9eeb-1cc2cc7c533a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.136036] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46da4206-a207-4ab3-b66e-fa7a0aa52f69 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.149993] env[61906]: DEBUG nova.compute.provider_tree [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.248201] env[61906]: INFO nova.compute.manager [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Took 32.33 seconds to build instance. [ 830.316170] env[61906]: DEBUG nova.compute.manager [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 830.333580] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.333803] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.334050] env[61906]: DEBUG nova.network.neutron [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 830.340025] env[61906]: DEBUG oslo_vmware.api [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356640, 'name': PowerOnVM_Task, 'duration_secs': 0.431354} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.340025] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 830.340025] env[61906]: DEBUG nova.compute.manager [None req-0e5ef164-243a-496e-bd4f-f660a322b88a tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.340684] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bed4ae1-b781-410d-b169-73cef4a2eb4a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.468981] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356641, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.562661] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356642, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.653612] env[61906]: DEBUG nova.scheduler.client.report [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.750249] env[61906]: DEBUG oslo_concurrency.lockutils [None req-439ed38a-2d4d-48c7-ae7d-81d17d944cfb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.333s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.866935] env[61906]: DEBUG nova.network.neutron [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.969475] env[61906]: DEBUG oslo_vmware.api [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356641, 'name': PowerOnVM_Task, 'duration_secs': 0.680747} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.969864] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 830.970352] env[61906]: INFO nova.compute.manager [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Took 8.47 seconds to spawn the instance on the hypervisor. [ 830.970599] env[61906]: DEBUG nova.compute.manager [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.971437] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b6abaa-7c02-40ad-871a-5f89d90d3e3b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.061322] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356642, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56343} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.061646] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 24b4a747-67aa-4388-aed6-cb646cd55765/24b4a747-67aa-4388-aed6-cb646cd55765.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 831.061974] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.062243] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b02e8bbc-15ff-441d-8855-44681b85d0a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.068674] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 831.068674] env[61906]: value = "task-1356643" [ 831.068674] env[61906]: _type = "Task" [ 831.068674] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.076771] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356643, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.101587] env[61906]: DEBUG nova.network.neutron [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updating instance_info_cache with network_info: [{"id": "f119e802-0a46-4390-927d-805c2e23a4b4", "address": "fa:16:3e:2c:b5:5c", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119e802-0a", "ovs_interfaceid": "f119e802-0a46-4390-927d-805c2e23a4b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.163930] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.855s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.166582] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.998s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.166582] env[61906]: DEBUG nova.objects.instance [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Lazy-loading 'resources' on Instance uuid 8ab6b473-d13c-4341-9789-992ac3aba6a2 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.199151] env[61906]: INFO nova.scheduler.client.report [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Deleted allocations for instance 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc [ 831.329348] env[61906]: DEBUG nova.compute.manager [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 831.390574] env[61906]: DEBUG nova.virt.hardware [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.390743] env[61906]: DEBUG nova.virt.hardware [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.390863] env[61906]: DEBUG nova.virt.hardware [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.391061] env[61906]: DEBUG nova.virt.hardware [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.391211] env[61906]: DEBUG nova.virt.hardware [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.391361] env[61906]: DEBUG nova.virt.hardware [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.393123] env[61906]: DEBUG nova.virt.hardware [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.393123] env[61906]: DEBUG nova.virt.hardware [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.393123] env[61906]: DEBUG nova.virt.hardware [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.393123] env[61906]: DEBUG nova.virt.hardware [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.393123] env[61906]: DEBUG nova.virt.hardware [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.393936] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb930982-b9c2-4603-a7fc-371becfb39c9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.403495] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ecfb2c-54ee-4574-b444-4b05d1727d2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.420153] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.427830] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Creating folder: Project (15830a6b7c1f4076905f9267b9cc80f8). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 831.427830] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f62a047f-e53f-45e6-b5e1-bbb43f59346e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.436392] env[61906]: DEBUG nova.compute.manager [req-07337218-fba0-4f8a-9ccf-d6b23eb48418 req-4e179005-4729-40da-9b2c-2618c3223e57 service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Received event network-changed-d803c8e8-7145-412c-aa3e-2545666742eb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.436392] env[61906]: DEBUG nova.compute.manager [req-07337218-fba0-4f8a-9ccf-d6b23eb48418 req-4e179005-4729-40da-9b2c-2618c3223e57 service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Refreshing instance network info cache due to event network-changed-d803c8e8-7145-412c-aa3e-2545666742eb. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.436392] env[61906]: DEBUG oslo_concurrency.lockutils [req-07337218-fba0-4f8a-9ccf-d6b23eb48418 req-4e179005-4729-40da-9b2c-2618c3223e57 service nova] Acquiring lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.436392] env[61906]: DEBUG oslo_concurrency.lockutils [req-07337218-fba0-4f8a-9ccf-d6b23eb48418 req-4e179005-4729-40da-9b2c-2618c3223e57 service nova] Acquired lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.436392] env[61906]: DEBUG nova.network.neutron [req-07337218-fba0-4f8a-9ccf-d6b23eb48418 req-4e179005-4729-40da-9b2c-2618c3223e57 service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Refreshing network info cache for port d803c8e8-7145-412c-aa3e-2545666742eb {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 831.438956] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Created folder: Project (15830a6b7c1f4076905f9267b9cc80f8) in parent group-v288914. [ 831.439155] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Creating folder: Instances. Parent ref: group-v288974. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 831.439584] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-20b8f28a-4253-46c0-a9ea-4a5004e515a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.454222] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Created folder: Instances in parent group-v288974. [ 831.454222] env[61906]: DEBUG oslo.service.loopingcall [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.454419] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 831.454769] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d59c818b-5125-4f84-bce6-689e1aca9e42 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.475065] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 831.475065] env[61906]: value = "task-1356646" [ 831.475065] env[61906]: _type = "Task" [ 831.475065] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.489224] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356646, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.493786] env[61906]: INFO nova.compute.manager [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Took 32.93 seconds to build instance. [ 831.580268] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356643, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065984} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.580683] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.581691] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606d7389-41e8-4dea-9ef3-c531b85b5974 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.608838] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 24b4a747-67aa-4388-aed6-cb646cd55765/24b4a747-67aa-4388-aed6-cb646cd55765.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 831.609403] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.609699] env[61906]: DEBUG nova.compute.manager [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Instance network_info: |[{"id": "f119e802-0a46-4390-927d-805c2e23a4b4", "address": "fa:16:3e:2c:b5:5c", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119e802-0a", "ovs_interfaceid": "f119e802-0a46-4390-927d-805c2e23a4b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 831.609990] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6bb7ceb4-1f32-4c0c-9e39-53d9343b58b7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.633261] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:b5:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8edfde4-5a99-4745-956d-04da82ab1b85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f119e802-0a46-4390-927d-805c2e23a4b4', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.645286] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating folder: Project (8f3bb9ce83b3430bb9202d3c72e77b4c). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 831.646496] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d4bc377c-57f5-4543-8d57-d1e78576934d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.653375] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 831.653375] env[61906]: value = "task-1356648" [ 831.653375] env[61906]: _type = "Task" [ 831.653375] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.658399] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Created folder: Project (8f3bb9ce83b3430bb9202d3c72e77b4c) in parent group-v288914. [ 831.658580] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating folder: Instances. Parent ref: group-v288977. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 831.659166] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a9ae3c20-beb0-4604-a711-337719dad44f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.664026] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356648, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.671976] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Created folder: Instances in parent group-v288977. [ 831.672333] env[61906]: DEBUG oslo.service.loopingcall [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.675640] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 831.675640] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-556e458f-9ada-42ee-9d71-66a574ec2f91 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.694328] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 831.694328] env[61906]: value = "task-1356650" [ 831.694328] env[61906]: _type = "Task" [ 831.694328] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.710519] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356650, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.711135] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5ce59871-2488-4128-af15-512053713be6 tempest-ServersTestMultiNic-1049742597 tempest-ServersTestMultiNic-1049742597-project-member] Lock "2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.383s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.848011] env[61906]: DEBUG nova.compute.manager [req-73d6fa48-8efd-4315-9948-4e99695c493a req-16f9dcd5-3523-47b4-bdb7-d1696b51375e service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Received event network-changed-f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.848731] env[61906]: DEBUG nova.compute.manager [req-73d6fa48-8efd-4315-9948-4e99695c493a req-16f9dcd5-3523-47b4-bdb7-d1696b51375e service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Refreshing instance network info cache due to event network-changed-f119e802-0a46-4390-927d-805c2e23a4b4. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.848731] env[61906]: DEBUG oslo_concurrency.lockutils [req-73d6fa48-8efd-4315-9948-4e99695c493a req-16f9dcd5-3523-47b4-bdb7-d1696b51375e service nova] Acquiring lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.848731] env[61906]: DEBUG oslo_concurrency.lockutils [req-73d6fa48-8efd-4315-9948-4e99695c493a req-16f9dcd5-3523-47b4-bdb7-d1696b51375e service nova] Acquired lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.848731] env[61906]: DEBUG nova.network.neutron [req-73d6fa48-8efd-4315-9948-4e99695c493a req-16f9dcd5-3523-47b4-bdb7-d1696b51375e service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Refreshing network info cache for port f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 831.915855] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e4cf85-0d27-4b21-8f21-63778f560d28 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.924845] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dfbd917-27b1-4890-af7c-03b6835e49d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.962912] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac6217f-8c4a-4b21-b2cb-f3894fdf3fff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.972206] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a1f9141-28de-456e-bfdf-c6b259f49279 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.994413] env[61906]: DEBUG nova.compute.provider_tree [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.000483] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3a33ec71-0094-461d-bba7-575c7f2b3820 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Lock "0552e012-edcf-444c-a9b3-f1cea93b9e82" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.327s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.001847] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356646, 'name': CreateVM_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.164434] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356648, 'name': ReconfigVM_Task, 'duration_secs': 0.505814} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.164434] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 24b4a747-67aa-4388-aed6-cb646cd55765/24b4a747-67aa-4388-aed6-cb646cd55765.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 832.164794] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-833268f6-b670-4ed7-914e-4bf813429e0d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.172158] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 832.172158] env[61906]: value = "task-1356651" [ 832.172158] env[61906]: _type = "Task" [ 832.172158] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.186022] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356651, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.203690] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356650, 'name': CreateVM_Task, 'duration_secs': 0.466674} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.203801] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 832.205325] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.205325] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.205325] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 832.205325] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e76f8bf7-bf45-4d92-acd6-5ae3d5050b19 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.211541] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 832.211541] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5294febd-b413-8468-9a33-72513d87bea2" [ 832.211541] env[61906]: _type = "Task" [ 832.211541] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.221518] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5294febd-b413-8468-9a33-72513d87bea2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.490940] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356646, 'name': CreateVM_Task, 'duration_secs': 0.814177} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.491838] env[61906]: DEBUG nova.network.neutron [req-07337218-fba0-4f8a-9ccf-d6b23eb48418 req-4e179005-4729-40da-9b2c-2618c3223e57 service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updated VIF entry in instance network info cache for port d803c8e8-7145-412c-aa3e-2545666742eb. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 832.492154] env[61906]: DEBUG nova.network.neutron [req-07337218-fba0-4f8a-9ccf-d6b23eb48418 req-4e179005-4729-40da-9b2c-2618c3223e57 service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance_info_cache with network_info: [{"id": "d803c8e8-7145-412c-aa3e-2545666742eb", "address": "fa:16:3e:46:59:77", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd803c8e8-71", "ovs_interfaceid": "d803c8e8-7145-412c-aa3e-2545666742eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.493285] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 832.494029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.505436] env[61906]: DEBUG nova.scheduler.client.report [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.668628] env[61906]: DEBUG nova.network.neutron [req-73d6fa48-8efd-4315-9948-4e99695c493a req-16f9dcd5-3523-47b4-bdb7-d1696b51375e service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updated VIF entry in instance network info cache for port f119e802-0a46-4390-927d-805c2e23a4b4. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 832.668628] env[61906]: DEBUG nova.network.neutron [req-73d6fa48-8efd-4315-9948-4e99695c493a req-16f9dcd5-3523-47b4-bdb7-d1696b51375e service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updating instance_info_cache with network_info: [{"id": "f119e802-0a46-4390-927d-805c2e23a4b4", "address": "fa:16:3e:2c:b5:5c", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119e802-0a", "ovs_interfaceid": "f119e802-0a46-4390-927d-805c2e23a4b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.682534] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356651, 'name': Rename_Task, 'duration_secs': 0.15207} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.683055] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 832.685260] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0f56929-19e1-44de-bef7-9351d5035b57 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.693027] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 832.693027] env[61906]: value = "task-1356652" [ 832.693027] env[61906]: _type = "Task" [ 832.693027] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.699626] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356652, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.725022] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5294febd-b413-8468-9a33-72513d87bea2, 'name': SearchDatastore_Task, 'duration_secs': 0.018334} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.725022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.725022] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.725022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.725022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.725022] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.725022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.725022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 832.725022] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64901b07-04ad-4414-92be-ae4f6284f99d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.726966] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65486424-a9cd-4f32-8552-7babcf29f473 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.734024] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 832.734024] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fbcd71-2dd8-09d6-6a35-4766fb618ee8" [ 832.734024] env[61906]: _type = "Task" [ 832.734024] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.742023] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.742023] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 832.742023] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf9dc7a4-7be4-4f03-88e7-d01d368cde03 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.744879] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fbcd71-2dd8-09d6-6a35-4766fb618ee8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.750179] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 832.750179] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5211d992-2b73-f8cf-18d5-0cf2b8b8dbad" [ 832.750179] env[61906]: _type = "Task" [ 832.750179] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.756725] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5211d992-2b73-f8cf-18d5-0cf2b8b8dbad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.996853] env[61906]: DEBUG oslo_concurrency.lockutils [req-07337218-fba0-4f8a-9ccf-d6b23eb48418 req-4e179005-4729-40da-9b2c-2618c3223e57 service nova] Releasing lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.015080] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.019548] env[61906]: DEBUG oslo_concurrency.lockutils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.539s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.021824] env[61906]: INFO nova.compute.claims [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 833.044842] env[61906]: INFO nova.scheduler.client.report [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Deleted allocations for instance 8ab6b473-d13c-4341-9789-992ac3aba6a2 [ 833.171140] env[61906]: DEBUG oslo_concurrency.lockutils [req-73d6fa48-8efd-4315-9948-4e99695c493a req-16f9dcd5-3523-47b4-bdb7-d1696b51375e service nova] Releasing lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.205084] env[61906]: DEBUG oslo_vmware.api [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356652, 'name': PowerOnVM_Task, 'duration_secs': 0.49604} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.206000] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.206332] env[61906]: INFO nova.compute.manager [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Took 8.40 seconds to spawn the instance on the hypervisor. [ 833.206835] env[61906]: DEBUG nova.compute.manager [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.207811] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a84f54-8398-4a5e-a263-2e3dc926bd16 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.244053] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fbcd71-2dd8-09d6-6a35-4766fb618ee8, 'name': SearchDatastore_Task, 'duration_secs': 0.011793} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.245352] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.245785] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 833.246145] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.259191] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5211d992-2b73-f8cf-18d5-0cf2b8b8dbad, 'name': SearchDatastore_Task, 'duration_secs': 0.025167} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.260456] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a45c6312-64d3-42ad-9ca8-4507ad569ddc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.268174] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 833.268174] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d2a875-2e36-9a7a-1d8f-bca2fbe73f67" [ 833.268174] env[61906]: _type = "Task" [ 833.268174] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.279990] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d2a875-2e36-9a7a-1d8f-bca2fbe73f67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.511232] env[61906]: INFO nova.compute.manager [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Rescuing [ 833.511232] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.511232] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.511232] env[61906]: DEBUG nova.network.neutron [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.553908] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d751840a-d44b-4284-bc76-edaf337e1001 tempest-ServerMetadataTestJSON-893466564 tempest-ServerMetadataTestJSON-893466564-project-member] Lock "8ab6b473-d13c-4341-9789-992ac3aba6a2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.553s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.591230] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "63cbf664-078b-40c5-92bb-7a5a8d8aea3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.591710] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "63cbf664-078b-40c5-92bb-7a5a8d8aea3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.624700] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "3a6da3bd-8e28-4c23-8660-f32f5c862d02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.624700] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "3a6da3bd-8e28-4c23-8660-f32f5c862d02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.732307] env[61906]: INFO nova.compute.manager [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Took 27.75 seconds to build instance. [ 833.782363] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d2a875-2e36-9a7a-1d8f-bca2fbe73f67, 'name': SearchDatastore_Task, 'duration_secs': 0.010805} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.786742] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.787801] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] cb879dc0-3af7-4279-aa28-66e2b8b4286d/cb879dc0-3af7-4279-aa28-66e2b8b4286d.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 833.792185] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.794580] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 833.794580] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9eae3323-e990-4140-8181-85d1c0d0f9fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.797018] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe2d4a91-dd3f-4e1a-85ac-6c4242aa8f89 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.804245] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 833.804245] env[61906]: value = "task-1356653" [ 833.804245] env[61906]: _type = "Task" [ 833.804245] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.811421] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 833.811794] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 833.813416] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-641f2713-14fd-4a01-a671-596c6b8f2a32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.820277] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356653, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.824021] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 833.824021] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]526f0e74-4362-61f2-9734-9a4c0301a304" [ 833.824021] env[61906]: _type = "Task" [ 833.824021] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.834051] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]526f0e74-4362-61f2-9734-9a4c0301a304, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.939357] env[61906]: DEBUG nova.compute.manager [req-52f98105-3a7b-4749-9cf0-a18400931768 req-f832de24-b117-4dbe-8de0-7dbf7ae45b9b service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Received event network-changed-96711ac0-bf70-468e-9ecc-70a5313e4bee {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.939556] env[61906]: DEBUG nova.compute.manager [req-52f98105-3a7b-4749-9cf0-a18400931768 req-f832de24-b117-4dbe-8de0-7dbf7ae45b9b service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Refreshing instance network info cache due to event network-changed-96711ac0-bf70-468e-9ecc-70a5313e4bee. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.939851] env[61906]: DEBUG oslo_concurrency.lockutils [req-52f98105-3a7b-4749-9cf0-a18400931768 req-f832de24-b117-4dbe-8de0-7dbf7ae45b9b service nova] Acquiring lock "refresh_cache-0552e012-edcf-444c-a9b3-f1cea93b9e82" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.940017] env[61906]: DEBUG oslo_concurrency.lockutils [req-52f98105-3a7b-4749-9cf0-a18400931768 req-f832de24-b117-4dbe-8de0-7dbf7ae45b9b service nova] Acquired lock "refresh_cache-0552e012-edcf-444c-a9b3-f1cea93b9e82" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.940342] env[61906]: DEBUG nova.network.neutron [req-52f98105-3a7b-4749-9cf0-a18400931768 req-f832de24-b117-4dbe-8de0-7dbf7ae45b9b service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Refreshing network info cache for port 96711ac0-bf70-468e-9ecc-70a5313e4bee {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 834.096193] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 834.126073] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 834.242140] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d247c32c-4c8f-4621-ac14-38d3f22a0d9a tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "24b4a747-67aa-4388-aed6-cb646cd55765" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.934s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.294399] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057a9134-711e-4801-a2cc-d0bf65d16ac7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.302205] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7f26b847-2b03-466f-b78e-e21f4d01d492 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Suspending the VM {{(pid=61906) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 834.302696] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-d82d3614-49c9-4032-9155-1d253d5511b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.315284] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8275aa39-b6b0-4ebe-aa6f-6a1534e986f4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.320131] env[61906]: DEBUG oslo_vmware.api [None req-7f26b847-2b03-466f-b78e-e21f4d01d492 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 834.320131] env[61906]: value = "task-1356654" [ 834.320131] env[61906]: _type = "Task" [ 834.320131] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.327999] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356653, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.340830] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf8ca87-1f60-4c95-8c83-823e0595c29c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.344667] env[61906]: DEBUG oslo_vmware.api [None req-7f26b847-2b03-466f-b78e-e21f4d01d492 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356654, 'name': SuspendVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.353520] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]526f0e74-4362-61f2-9734-9a4c0301a304, 'name': SearchDatastore_Task, 'duration_secs': 0.015929} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.384907] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14c746c0-2d98-4400-b6dc-aca3447981b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.387872] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3497e4-d2d2-43b7-b5f1-ebd24179c8eb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.398370] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 834.398370] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5248a0fa-b305-93fc-fb03-802a7e3ff93b" [ 834.398370] env[61906]: _type = "Task" [ 834.398370] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.399797] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757f49f2-ff49-4991-9cd1-6366edb5a1db {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.423685] env[61906]: DEBUG nova.compute.provider_tree [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.429381] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5248a0fa-b305-93fc-fb03-802a7e3ff93b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.568793] env[61906]: DEBUG nova.network.neutron [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Updating instance_info_cache with network_info: [{"id": "e291dedf-5c3f-4d78-bad3-295b2d1e2f75", "address": "fa:16:3e:c8:8d:ef", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape291dedf-5c", "ovs_interfaceid": "e291dedf-5c3f-4d78-bad3-295b2d1e2f75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.630756] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.654373] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.816647] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356653, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.875441} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.816936] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] cb879dc0-3af7-4279-aa28-66e2b8b4286d/cb879dc0-3af7-4279-aa28-66e2b8b4286d.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 834.817146] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.817489] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aded43de-98d7-4479-9a31-1803ec691bce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.828753] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 834.828753] env[61906]: value = "task-1356655" [ 834.828753] env[61906]: _type = "Task" [ 834.828753] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.835468] env[61906]: DEBUG oslo_vmware.api [None req-7f26b847-2b03-466f-b78e-e21f4d01d492 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356654, 'name': SuspendVM_Task} progress is 79%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.840882] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356655, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.917191] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5248a0fa-b305-93fc-fb03-802a7e3ff93b, 'name': SearchDatastore_Task, 'duration_secs': 0.058857} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.917191] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.917191] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] a5f780a2-0cb1-4da8-8276-82b9653bf6a7/a5f780a2-0cb1-4da8-8276-82b9653bf6a7.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 834.917457] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3eb0d498-cd97-43a2-8ce6-7dbb17bac2b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.924572] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 834.924572] env[61906]: value = "task-1356656" [ 834.924572] env[61906]: _type = "Task" [ 834.924572] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.931575] env[61906]: DEBUG nova.scheduler.client.report [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.940279] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356656, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.034018] env[61906]: DEBUG nova.network.neutron [req-52f98105-3a7b-4749-9cf0-a18400931768 req-f832de24-b117-4dbe-8de0-7dbf7ae45b9b service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Updated VIF entry in instance network info cache for port 96711ac0-bf70-468e-9ecc-70a5313e4bee. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 835.034457] env[61906]: DEBUG nova.network.neutron [req-52f98105-3a7b-4749-9cf0-a18400931768 req-f832de24-b117-4dbe-8de0-7dbf7ae45b9b service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Updating instance_info_cache with network_info: [{"id": "96711ac0-bf70-468e-9ecc-70a5313e4bee", "address": "fa:16:3e:64:a0:62", "network": {"id": "65476ec7-1ddc-4b0f-86bc-090782193282", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-990976432-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.252", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ec67c1b35414992a70b6dfa372ac12e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496faa4d-d874-449b-905e-328ddd60b31b", "external-id": "nsx-vlan-transportzone-391", "segmentation_id": 391, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96711ac0-bf", "ovs_interfaceid": "96711ac0-bf70-468e-9ecc-70a5313e4bee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.075018] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.333707] env[61906]: DEBUG oslo_vmware.api [None req-7f26b847-2b03-466f-b78e-e21f4d01d492 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356654, 'name': SuspendVM_Task, 'duration_secs': 0.815354} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.337643] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7f26b847-2b03-466f-b78e-e21f4d01d492 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Suspended the VM {{(pid=61906) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 835.337904] env[61906]: DEBUG nova.compute.manager [None req-7f26b847-2b03-466f-b78e-e21f4d01d492 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 835.338801] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a234aa7-4104-44a4-a5b5-c6bba379d0f4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.347639] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356655, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080821} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.349564] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.353779] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab977dbe-4ae8-498f-b0aa-3982cff1eb56 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.392127] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] cb879dc0-3af7-4279-aa28-66e2b8b4286d/cb879dc0-3af7-4279-aa28-66e2b8b4286d.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.392463] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66372fd4-ef71-451e-9a4e-83dd9233dd2b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.424028] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 835.424028] env[61906]: value = "task-1356657" [ 835.424028] env[61906]: _type = "Task" [ 835.424028] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.435154] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356657, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.436813] env[61906]: DEBUG oslo_concurrency.lockutils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.439017] env[61906]: DEBUG nova.compute.manager [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 835.446024] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.740s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.446024] env[61906]: INFO nova.compute.claims [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 835.448392] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356656, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.537361] env[61906]: DEBUG oslo_concurrency.lockutils [req-52f98105-3a7b-4749-9cf0-a18400931768 req-f832de24-b117-4dbe-8de0-7dbf7ae45b9b service nova] Releasing lock "refresh_cache-0552e012-edcf-444c-a9b3-f1cea93b9e82" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.608220] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 835.608220] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-867c0108-1772-4e9b-9890-ad8892cc1e92 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.615771] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 835.615771] env[61906]: value = "task-1356658" [ 835.615771] env[61906]: _type = "Task" [ 835.615771] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.624594] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356658, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.940371] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356657, 'name': ReconfigVM_Task, 'duration_secs': 0.279889} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.941996] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Reconfigured VM instance instance-00000049 to attach disk [datastore2] cb879dc0-3af7-4279-aa28-66e2b8b4286d/cb879dc0-3af7-4279-aa28-66e2b8b4286d.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.941996] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b820416f-35c9-4d45-b8a5-76ea5b2743e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.947431] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356656, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.559307} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.948080] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] a5f780a2-0cb1-4da8-8276-82b9653bf6a7/a5f780a2-0cb1-4da8-8276-82b9653bf6a7.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 835.948344] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.948623] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-02c7c218-a83e-47c2-811c-92a98e7ac489 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.951380] env[61906]: DEBUG nova.compute.utils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 835.955951] env[61906]: DEBUG nova.compute.manager [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 835.956736] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 835.956736] env[61906]: value = "task-1356659" [ 835.956736] env[61906]: _type = "Task" [ 835.956736] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.961804] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 835.961804] env[61906]: value = "task-1356660" [ 835.961804] env[61906]: _type = "Task" [ 835.961804] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.970778] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356659, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.977049] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356660, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.133876] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356658, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.459174] env[61906]: DEBUG nova.compute.manager [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 836.489916] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356659, 'name': Rename_Task, 'duration_secs': 0.350129} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.495254] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 836.495254] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356660, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.132402} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.495254] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e95904ef-dc1b-493e-8fea-7b5edafc2fbd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.496372] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.499044] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-446e5b4a-95d9-4e7f-ad09-ee18fa0c1d49 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.535282] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] a5f780a2-0cb1-4da8-8276-82b9653bf6a7/a5f780a2-0cb1-4da8-8276-82b9653bf6a7.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.541692] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-09280260-412b-42c5-953d-b8481d3f9ed9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.568106] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 836.568106] env[61906]: value = "task-1356661" [ 836.568106] env[61906]: _type = "Task" [ 836.568106] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.579239] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 836.579239] env[61906]: value = "task-1356662" [ 836.579239] env[61906]: _type = "Task" [ 836.579239] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.594030] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356661, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.601389] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356662, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.631706] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356658, 'name': PowerOffVM_Task, 'duration_secs': 0.518615} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.631706] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 836.631706] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f61448-51b6-4c48-b259-cb9c8db1c8fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.652243] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6185275-4596-448f-a20a-199cbcc53073 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.696311] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 836.697563] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71426094-798a-4286-868e-6dce294eae75 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.705234] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 836.705234] env[61906]: value = "task-1356663" [ 836.705234] env[61906]: _type = "Task" [ 836.705234] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.715682] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] VM already powered off {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 836.715767] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.716111] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.716213] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.716376] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.716662] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dc1b0bed-d5b2-4861-a6b1-97e8d4d11d52 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.726094] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.726279] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 836.727052] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4d3c765-18f3-4793-b98e-844f54c18f19 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.732421] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 836.732421] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522aceaa-8e1e-ceaa-ec68-72915397b3c2" [ 836.732421] env[61906]: _type = "Task" [ 836.732421] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.742762] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522aceaa-8e1e-ceaa-ec68-72915397b3c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.818122] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d22964-9936-459f-ac6d-3bbb4f9fafb4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.824735] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60bff834-9bf9-4abe-85d9-4a4c3bef4d24 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.862626] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18c23c9-a8c5-4d4c-ac86-b327e3f953ed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.874331] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e492fe5f-25c4-43ac-bd90-be9357dc74b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.891297] env[61906]: DEBUG nova.compute.provider_tree [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.081633] env[61906]: DEBUG oslo_vmware.api [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356661, 'name': PowerOnVM_Task, 'duration_secs': 0.465571} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.082524] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 837.082865] env[61906]: INFO nova.compute.manager [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Took 8.17 seconds to spawn the instance on the hypervisor. [ 837.083189] env[61906]: DEBUG nova.compute.manager [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.084606] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67edf0ed-c7a5-4ee9-ab96-2f6708e71c3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.096723] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356662, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.243020] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522aceaa-8e1e-ceaa-ec68-72915397b3c2, 'name': SearchDatastore_Task, 'duration_secs': 0.028516} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.243848] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cbe1726-0428-4431-8927-33f65cb56385 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.250261] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 837.250261] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52603134-3207-bab5-a512-26ba4cf7dece" [ 837.250261] env[61906]: _type = "Task" [ 837.250261] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.258614] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52603134-3207-bab5-a512-26ba4cf7dece, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.395082] env[61906]: DEBUG nova.scheduler.client.report [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 837.476184] env[61906]: DEBUG nova.compute.manager [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 837.509069] env[61906]: DEBUG nova.virt.hardware [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 837.509069] env[61906]: DEBUG nova.virt.hardware [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 837.509069] env[61906]: DEBUG nova.virt.hardware [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 837.509069] env[61906]: DEBUG nova.virt.hardware [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 837.509069] env[61906]: DEBUG nova.virt.hardware [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 837.509069] env[61906]: DEBUG nova.virt.hardware [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 837.509608] env[61906]: DEBUG nova.virt.hardware [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 837.509966] env[61906]: DEBUG nova.virt.hardware [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 837.510336] env[61906]: DEBUG nova.virt.hardware [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 837.510717] env[61906]: DEBUG nova.virt.hardware [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 837.511085] env[61906]: DEBUG nova.virt.hardware [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.514129] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b4e5c31-cfc2-4fd1-9338-a5188608a86d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.523479] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-118d249c-9875-48c1-9dd0-145400ffdf98 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.530792] env[61906]: DEBUG nova.compute.manager [req-30d555e5-8d1d-4dc9-8af4-2efef1fa64ec req-55c95480-5a26-46c4-974d-27de7c74173d service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Received event network-changed-96711ac0-bf70-468e-9ecc-70a5313e4bee {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.531171] env[61906]: DEBUG nova.compute.manager [req-30d555e5-8d1d-4dc9-8af4-2efef1fa64ec req-55c95480-5a26-46c4-974d-27de7c74173d service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Refreshing instance network info cache due to event network-changed-96711ac0-bf70-468e-9ecc-70a5313e4bee. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 837.531521] env[61906]: DEBUG oslo_concurrency.lockutils [req-30d555e5-8d1d-4dc9-8af4-2efef1fa64ec req-55c95480-5a26-46c4-974d-27de7c74173d service nova] Acquiring lock "refresh_cache-0552e012-edcf-444c-a9b3-f1cea93b9e82" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.533046] env[61906]: DEBUG oslo_concurrency.lockutils [req-30d555e5-8d1d-4dc9-8af4-2efef1fa64ec req-55c95480-5a26-46c4-974d-27de7c74173d service nova] Acquired lock "refresh_cache-0552e012-edcf-444c-a9b3-f1cea93b9e82" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.533046] env[61906]: DEBUG nova.network.neutron [req-30d555e5-8d1d-4dc9-8af4-2efef1fa64ec req-55c95480-5a26-46c4-974d-27de7c74173d service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Refreshing network info cache for port 96711ac0-bf70-468e-9ecc-70a5313e4bee {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 837.547545] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.554088] env[61906]: DEBUG oslo.service.loopingcall [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.555331] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.555703] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-41bcfe1d-9778-4e47-94a8-97913b08200b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.574292] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.574292] env[61906]: value = "task-1356664" [ 837.574292] env[61906]: _type = "Task" [ 837.574292] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.583845] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356664, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.593041] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356662, 'name': ReconfigVM_Task, 'duration_secs': 0.743136} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.593361] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Reconfigured VM instance instance-0000004a to attach disk [datastore2] a5f780a2-0cb1-4da8-8276-82b9653bf6a7/a5f780a2-0cb1-4da8-8276-82b9653bf6a7.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.593967] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b2c911f8-eeff-4c49-b2d1-cfe8ecbba2d9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.600034] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 837.600034] env[61906]: value = "task-1356665" [ 837.600034] env[61906]: _type = "Task" [ 837.600034] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.610776] env[61906]: INFO nova.compute.manager [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Took 29.56 seconds to build instance. [ 837.616032] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356665, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.762410] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52603134-3207-bab5-a512-26ba4cf7dece, 'name': SearchDatastore_Task, 'duration_secs': 0.010217} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.762410] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.762410] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 79452791-59cb-4722-bb4a-8e59d8c4e641/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1-rescue.vmdk. {{(pid=61906) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 837.762838] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e465e4e-4112-4bdb-b538-1857e9186296 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.770529] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 837.770529] env[61906]: value = "task-1356666" [ 837.770529] env[61906]: _type = "Task" [ 837.770529] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.781132] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356666, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.879984] env[61906]: DEBUG nova.compute.manager [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.881857] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee670c3-c925-4cda-9b10-11eee07c6c8b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.902841] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.903384] env[61906]: DEBUG nova.compute.manager [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 837.905997] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.493s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.907808] env[61906]: INFO nova.compute.claims [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.087999] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356664, 'name': CreateVM_Task, 'duration_secs': 0.360978} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.088191] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 838.088614] env[61906]: DEBUG oslo_concurrency.lockutils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.088791] env[61906]: DEBUG oslo_concurrency.lockutils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.089151] env[61906]: DEBUG oslo_concurrency.lockutils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.089408] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afabff3f-00ad-449f-b772-ea7119a9b962 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.094436] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 838.094436] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527a5dd9-4e15-735a-294b-728dbca94be1" [ 838.094436] env[61906]: _type = "Task" [ 838.094436] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.103734] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527a5dd9-4e15-735a-294b-728dbca94be1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.115365] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356665, 'name': Rename_Task, 'duration_secs': 0.178937} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.115365] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 838.115365] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-adfe4e82-e7d3-463d-b3f2-93198d32d08c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.117530] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ac4da877-5e18-4df9-b45a-619be4f49201 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.788s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.120853] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 838.120853] env[61906]: value = "task-1356667" [ 838.120853] env[61906]: _type = "Task" [ 838.120853] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.130878] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356667, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.285251] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356666, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.398834] env[61906]: INFO nova.compute.manager [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] instance snapshotting [ 838.399040] env[61906]: WARNING nova.compute.manager [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 838.402353] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1671439b-05e6-40d3-8c5a-1779410f5003 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.428762] env[61906]: DEBUG nova.compute.utils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 838.431972] env[61906]: DEBUG nova.compute.manager [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 838.432228] env[61906]: DEBUG nova.network.neutron [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 838.434806] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ba90c3-db6c-43d3-b463-1fe4824bba4f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.509197] env[61906]: DEBUG nova.policy [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6c0cf3c5f106420e98c7fd53ac2d2733', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9da9bb542c6a4aba9e99b402d7f89577', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 838.581142] env[61906]: DEBUG nova.network.neutron [req-30d555e5-8d1d-4dc9-8af4-2efef1fa64ec req-55c95480-5a26-46c4-974d-27de7c74173d service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Updated VIF entry in instance network info cache for port 96711ac0-bf70-468e-9ecc-70a5313e4bee. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 838.581860] env[61906]: DEBUG nova.network.neutron [req-30d555e5-8d1d-4dc9-8af4-2efef1fa64ec req-55c95480-5a26-46c4-974d-27de7c74173d service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Updating instance_info_cache with network_info: [{"id": "96711ac0-bf70-468e-9ecc-70a5313e4bee", "address": "fa:16:3e:64:a0:62", "network": {"id": "65476ec7-1ddc-4b0f-86bc-090782193282", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-990976432-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ec67c1b35414992a70b6dfa372ac12e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496faa4d-d874-449b-905e-328ddd60b31b", "external-id": "nsx-vlan-transportzone-391", "segmentation_id": 391, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96711ac0-bf", "ovs_interfaceid": "96711ac0-bf70-468e-9ecc-70a5313e4bee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.608837] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527a5dd9-4e15-735a-294b-728dbca94be1, 'name': SearchDatastore_Task, 'duration_secs': 0.016552} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.609176] env[61906]: DEBUG oslo_concurrency.lockutils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.609411] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.609739] env[61906]: DEBUG oslo_concurrency.lockutils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.609931] env[61906]: DEBUG oslo_concurrency.lockutils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.610190] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.610543] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c9bade89-a2cf-4984-9abb-cc43299fe414 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.628229] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.628229] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 838.630131] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9abb5a7-f21f-408b-a755-1482811c0a5d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.638730] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356667, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.642710] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 838.642710] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52da414c-9713-ab23-7b93-6dab667e19dd" [ 838.642710] env[61906]: _type = "Task" [ 838.642710] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.654628] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52da414c-9713-ab23-7b93-6dab667e19dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.783371] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356666, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.64109} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.783371] env[61906]: INFO nova.virt.vmwareapi.ds_util [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 79452791-59cb-4722-bb4a-8e59d8c4e641/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1-rescue.vmdk. [ 838.783371] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c47ce8b-90b5-4db5-bd0a-fa1ba26fc454 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.809819] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 79452791-59cb-4722-bb4a-8e59d8c4e641/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.810789] env[61906]: DEBUG nova.network.neutron [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Successfully created port: fcbc3598-49e0-4da8-91e1-e0f6929f305f {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 838.812580] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1979a19-f16c-44e5-a070-efdd8e3d3536 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.831015] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 838.831015] env[61906]: value = "task-1356668" [ 838.831015] env[61906]: _type = "Task" [ 838.831015] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.839202] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356668, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.930907] env[61906]: DEBUG nova.compute.manager [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 838.956398] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 838.956798] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0b5ef31d-b213-4067-9004-433db8cdb359 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.969013] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 838.969013] env[61906]: value = "task-1356669" [ 838.969013] env[61906]: _type = "Task" [ 838.969013] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.978121] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356669, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.084632] env[61906]: DEBUG oslo_concurrency.lockutils [req-30d555e5-8d1d-4dc9-8af4-2efef1fa64ec req-55c95480-5a26-46c4-974d-27de7c74173d service nova] Releasing lock "refresh_cache-0552e012-edcf-444c-a9b3-f1cea93b9e82" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.118244] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Acquiring lock "0552e012-edcf-444c-a9b3-f1cea93b9e82" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.118539] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Lock "0552e012-edcf-444c-a9b3-f1cea93b9e82" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.118760] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Acquiring lock "0552e012-edcf-444c-a9b3-f1cea93b9e82-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.118949] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Lock "0552e012-edcf-444c-a9b3-f1cea93b9e82-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.119150] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Lock "0552e012-edcf-444c-a9b3-f1cea93b9e82-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.121422] env[61906]: INFO nova.compute.manager [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Terminating instance [ 839.128742] env[61906]: DEBUG nova.compute.manager [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 839.128896] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 839.129699] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad782fb2-60d1-4929-b976-98d95db0051c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.137632] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.140602] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31d9519e-a722-4260-bc00-e70cfd5f9bd4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.142381] env[61906]: DEBUG oslo_vmware.api [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356667, 'name': PowerOnVM_Task, 'duration_secs': 0.623457} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.142628] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 839.142824] env[61906]: INFO nova.compute.manager [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Took 7.81 seconds to spawn the instance on the hypervisor. [ 839.142999] env[61906]: DEBUG nova.compute.manager [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.147019] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e210eedf-d6db-4bd4-adec-d359499bc9f6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.150457] env[61906]: DEBUG oslo_vmware.api [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Waiting for the task: (returnval){ [ 839.150457] env[61906]: value = "task-1356670" [ 839.150457] env[61906]: _type = "Task" [ 839.150457] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.163602] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52da414c-9713-ab23-7b93-6dab667e19dd, 'name': SearchDatastore_Task, 'duration_secs': 0.03064} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.167810] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4203ad1b-91d3-4648-b588-e04fad5f6360 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.172491] env[61906]: DEBUG oslo_vmware.api [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356670, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.175168] env[61906]: DEBUG nova.compute.manager [req-c7b997e1-3423-470f-bb1f-ed7cc222c208 req-1ccf0fd6-a66e-4c46-a688-200cb51f677c service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Received event network-changed-f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 839.175600] env[61906]: DEBUG nova.compute.manager [req-c7b997e1-3423-470f-bb1f-ed7cc222c208 req-1ccf0fd6-a66e-4c46-a688-200cb51f677c service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Refreshing instance network info cache due to event network-changed-f119e802-0a46-4390-927d-805c2e23a4b4. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 839.175600] env[61906]: DEBUG oslo_concurrency.lockutils [req-c7b997e1-3423-470f-bb1f-ed7cc222c208 req-1ccf0fd6-a66e-4c46-a688-200cb51f677c service nova] Acquiring lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.175600] env[61906]: DEBUG oslo_concurrency.lockutils [req-c7b997e1-3423-470f-bb1f-ed7cc222c208 req-1ccf0fd6-a66e-4c46-a688-200cb51f677c service nova] Acquired lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.175834] env[61906]: DEBUG nova.network.neutron [req-c7b997e1-3423-470f-bb1f-ed7cc222c208 req-1ccf0fd6-a66e-4c46-a688-200cb51f677c service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Refreshing network info cache for port f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 839.179444] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 839.179444] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]520c848b-22cd-45e8-7b1e-c8d1b74f960a" [ 839.179444] env[61906]: _type = "Task" [ 839.179444] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.187948] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]520c848b-22cd-45e8-7b1e-c8d1b74f960a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.190128] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be4d0c6-8c62-4115-bf22-590bf30a465f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.198053] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa91153-c2fc-4bd2-b3f2-44b46a82bc50 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.237539] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25b4f08-5e94-45d6-8b8d-2e976d703f45 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.246828] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7797d95c-7848-4b4c-b0af-2425ddc73697 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.261726] env[61906]: DEBUG nova.compute.provider_tree [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.341843] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356668, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.478425] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356669, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.665061] env[61906]: DEBUG oslo_vmware.api [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356670, 'name': PowerOffVM_Task, 'duration_secs': 0.205921} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.665217] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 839.665371] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 839.665667] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ccce9884-a650-49e1-a12c-dfd73de3757f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.681610] env[61906]: INFO nova.compute.manager [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Took 31.19 seconds to build instance. [ 839.692111] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]520c848b-22cd-45e8-7b1e-c8d1b74f960a, 'name': SearchDatastore_Task, 'duration_secs': 0.018139} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.692383] env[61906]: DEBUG oslo_concurrency.lockutils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.692644] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44/31d0f640-47f4-4b72-b9fe-6ef4c76c6b44.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 839.693331] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7fdc2146-f094-4d19-9d58-09f928d4d3c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.700887] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 839.700887] env[61906]: value = "task-1356672" [ 839.700887] env[61906]: _type = "Task" [ 839.700887] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.710782] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356672, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.728452] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 839.730238] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 839.730238] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Deleting the datastore file [datastore1] 0552e012-edcf-444c-a9b3-f1cea93b9e82 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.730238] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e0842b0f-10b6-4270-8032-e82acca0e9f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.737138] env[61906]: DEBUG oslo_vmware.api [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Waiting for the task: (returnval){ [ 839.737138] env[61906]: value = "task-1356673" [ 839.737138] env[61906]: _type = "Task" [ 839.737138] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.748090] env[61906]: DEBUG oslo_vmware.api [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356673, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.765436] env[61906]: DEBUG nova.scheduler.client.report [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.841215] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356668, 'name': ReconfigVM_Task, 'duration_secs': 0.638357} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.841616] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 79452791-59cb-4722-bb4a-8e59d8c4e641/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.842759] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5d6a4e-1dbd-466e-a513-91a41296e6d3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.870913] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c9822c5-edfc-4efb-b5de-d2bee6d913a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.888233] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 839.888233] env[61906]: value = "task-1356674" [ 839.888233] env[61906]: _type = "Task" [ 839.888233] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.896433] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356674, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.916920] env[61906]: DEBUG nova.network.neutron [req-c7b997e1-3423-470f-bb1f-ed7cc222c208 req-1ccf0fd6-a66e-4c46-a688-200cb51f677c service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updated VIF entry in instance network info cache for port f119e802-0a46-4390-927d-805c2e23a4b4. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 839.917302] env[61906]: DEBUG nova.network.neutron [req-c7b997e1-3423-470f-bb1f-ed7cc222c208 req-1ccf0fd6-a66e-4c46-a688-200cb51f677c service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updating instance_info_cache with network_info: [{"id": "f119e802-0a46-4390-927d-805c2e23a4b4", "address": "fa:16:3e:2c:b5:5c", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119e802-0a", "ovs_interfaceid": "f119e802-0a46-4390-927d-805c2e23a4b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.940739] env[61906]: DEBUG nova.compute.manager [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 839.965044] env[61906]: DEBUG nova.virt.hardware [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 839.965357] env[61906]: DEBUG nova.virt.hardware [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 839.965559] env[61906]: DEBUG nova.virt.hardware [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.965842] env[61906]: DEBUG nova.virt.hardware [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 839.966064] env[61906]: DEBUG nova.virt.hardware [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.966284] env[61906]: DEBUG nova.virt.hardware [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 839.966569] env[61906]: DEBUG nova.virt.hardware [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 839.966765] env[61906]: DEBUG nova.virt.hardware [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 839.966989] env[61906]: DEBUG nova.virt.hardware [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 839.967230] env[61906]: DEBUG nova.virt.hardware [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 839.967462] env[61906]: DEBUG nova.virt.hardware [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.968787] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1578b305-898a-47ef-8789-ceb8840c662e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.983104] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50aad64-b251-499d-9818-d3bd21a53cee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.987676] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356669, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.187412] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c0f9aa5f-9a19-4957-ae1d-6410ee8c6fa2 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "a5f780a2-0cb1-4da8-8276-82b9653bf6a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.517s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.211361] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356672, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.216680] env[61906]: DEBUG nova.compute.manager [req-4db7c9e6-d99e-406b-ade9-1c4d4e6b215b req-dd7bc363-8c28-4b28-8643-1fb2f2b63fc8 service nova] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Received event network-vif-plugged-fcbc3598-49e0-4da8-91e1-e0f6929f305f {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.217076] env[61906]: DEBUG oslo_concurrency.lockutils [req-4db7c9e6-d99e-406b-ade9-1c4d4e6b215b req-dd7bc363-8c28-4b28-8643-1fb2f2b63fc8 service nova] Acquiring lock "0af2b07b-0632-4376-b502-58fe23d79f02-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.217181] env[61906]: DEBUG oslo_concurrency.lockutils [req-4db7c9e6-d99e-406b-ade9-1c4d4e6b215b req-dd7bc363-8c28-4b28-8643-1fb2f2b63fc8 service nova] Lock "0af2b07b-0632-4376-b502-58fe23d79f02-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.217364] env[61906]: DEBUG oslo_concurrency.lockutils [req-4db7c9e6-d99e-406b-ade9-1c4d4e6b215b req-dd7bc363-8c28-4b28-8643-1fb2f2b63fc8 service nova] Lock "0af2b07b-0632-4376-b502-58fe23d79f02-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.217547] env[61906]: DEBUG nova.compute.manager [req-4db7c9e6-d99e-406b-ade9-1c4d4e6b215b req-dd7bc363-8c28-4b28-8643-1fb2f2b63fc8 service nova] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] No waiting events found dispatching network-vif-plugged-fcbc3598-49e0-4da8-91e1-e0f6929f305f {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 840.217738] env[61906]: WARNING nova.compute.manager [req-4db7c9e6-d99e-406b-ade9-1c4d4e6b215b req-dd7bc363-8c28-4b28-8643-1fb2f2b63fc8 service nova] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Received unexpected event network-vif-plugged-fcbc3598-49e0-4da8-91e1-e0f6929f305f for instance with vm_state building and task_state spawning. [ 840.247756] env[61906]: DEBUG oslo_vmware.api [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356673, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.272377] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.272920] env[61906]: DEBUG nova.compute.manager [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 840.279550] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.163s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.279776] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.279941] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 840.280316] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.701s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.280524] env[61906]: DEBUG nova.objects.instance [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lazy-loading 'resources' on Instance uuid 9a23071d-40fd-4446-aa03-ecbddcafe3d6 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 840.290678] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57233105-0c1e-4187-9c1a-fb43c96bb5db {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.305245] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2589ad1b-71f3-4e74-b980-dc9714816fbd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.322340] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16d0299-ff0d-4c99-a9eb-4448ff28f59b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.332715] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3a6368-13af-4155-b1ee-af51bb932724 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.369021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181527MB free_disk=150GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 840.369194] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.400669] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356674, 'name': ReconfigVM_Task, 'duration_secs': 0.420364} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.401074] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.401377] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a5ca6bf-b44e-48ca-8d9f-041f88b5b7f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.410184] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 840.410184] env[61906]: value = "task-1356675" [ 840.410184] env[61906]: _type = "Task" [ 840.410184] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.421102] env[61906]: DEBUG oslo_concurrency.lockutils [req-c7b997e1-3423-470f-bb1f-ed7cc222c208 req-1ccf0fd6-a66e-4c46-a688-200cb51f677c service nova] Releasing lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.421591] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356675, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.481621] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356669, 'name': CreateSnapshot_Task, 'duration_secs': 1.07209} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.481888] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 840.482623] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe5793b-9f7f-4777-95c2-c1af44c32cdc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.711869] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356672, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.739669} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.712778] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44/31d0f640-47f4-4b72-b9fe-6ef4c76c6b44.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 840.713061] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.713335] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ececb99-e990-4315-b51c-aa288f7b1c88 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.720205] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 840.720205] env[61906]: value = "task-1356676" [ 840.720205] env[61906]: _type = "Task" [ 840.720205] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.728358] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356676, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.746176] env[61906]: DEBUG oslo_vmware.api [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Task: {'id': task-1356673, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.519113} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.746478] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.746766] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 840.747020] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 840.747253] env[61906]: INFO nova.compute.manager [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Took 1.62 seconds to destroy the instance on the hypervisor. [ 840.747544] env[61906]: DEBUG oslo.service.loopingcall [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.747762] env[61906]: DEBUG nova.compute.manager [-] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.747922] env[61906]: DEBUG nova.network.neutron [-] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 840.781162] env[61906]: DEBUG nova.compute.utils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.783038] env[61906]: DEBUG nova.compute.manager [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 840.783134] env[61906]: DEBUG nova.network.neutron [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 840.786127] env[61906]: DEBUG nova.network.neutron [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Successfully updated port: fcbc3598-49e0-4da8-91e1-e0f6929f305f {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.839673] env[61906]: DEBUG nova.compute.manager [req-41f117e3-c9a4-4d53-902c-05f0123ae196 req-b3294b28-7801-4de0-bd24-1b143d24748f service nova] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Received event network-changed-fcbc3598-49e0-4da8-91e1-e0f6929f305f {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.839875] env[61906]: DEBUG nova.compute.manager [req-41f117e3-c9a4-4d53-902c-05f0123ae196 req-b3294b28-7801-4de0-bd24-1b143d24748f service nova] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Refreshing instance network info cache due to event network-changed-fcbc3598-49e0-4da8-91e1-e0f6929f305f. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 840.840106] env[61906]: DEBUG oslo_concurrency.lockutils [req-41f117e3-c9a4-4d53-902c-05f0123ae196 req-b3294b28-7801-4de0-bd24-1b143d24748f service nova] Acquiring lock "refresh_cache-0af2b07b-0632-4376-b502-58fe23d79f02" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.840318] env[61906]: DEBUG oslo_concurrency.lockutils [req-41f117e3-c9a4-4d53-902c-05f0123ae196 req-b3294b28-7801-4de0-bd24-1b143d24748f service nova] Acquired lock "refresh_cache-0af2b07b-0632-4376-b502-58fe23d79f02" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.840406] env[61906]: DEBUG nova.network.neutron [req-41f117e3-c9a4-4d53-902c-05f0123ae196 req-b3294b28-7801-4de0-bd24-1b143d24748f service nova] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Refreshing network info cache for port fcbc3598-49e0-4da8-91e1-e0f6929f305f {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.907243] env[61906]: DEBUG nova.policy [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca80129d492e4d68b96ef14a63336ef3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb85e88fd4f54e0db2ff131f81137f64', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 840.920029] env[61906]: DEBUG oslo_vmware.api [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356675, 'name': PowerOnVM_Task, 'duration_secs': 0.470423} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.922193] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 840.925081] env[61906]: DEBUG nova.compute.manager [None req-3e7b9892-2714-413d-808f-a4dc71378638 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.925820] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb6da44-6169-42a0-8fda-4dce76a026e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.002854] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 841.003686] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-be52397e-d186-4a47-afe2-234c6eee56c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.012402] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 841.012402] env[61906]: value = "task-1356677" [ 841.012402] env[61906]: _type = "Task" [ 841.012402] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.024141] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356677, 'name': CloneVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.046304] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2215efbd-9b90-41f7-b863-1de312686ac7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.053673] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa056b4e-4698-4ded-9b38-d663a86cf19b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.083138] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faa6f00c-b607-4554-91d5-39c962d36375 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.090789] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a867e913-bcf0-481e-bd6f-f56931cb5b6e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.103870] env[61906]: DEBUG nova.compute.provider_tree [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.231250] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356676, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067487} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.231633] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.232664] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7c4ee9-3c2b-45f9-ad8c-642397755059 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.255500] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44/31d0f640-47f4-4b72-b9fe-6ef4c76c6b44.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.257898] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1126a5b9-e6bf-4762-bda0-498d2d92decc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.277112] env[61906]: DEBUG nova.compute.manager [req-402de077-dcd7-4ab9-b971-1ba51e5cca09 req-cfc51851-e0b1-41cf-83f8-02345f0c0588 service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Received event network-vif-deleted-96711ac0-bf70-468e-9ecc-70a5313e4bee {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.277338] env[61906]: INFO nova.compute.manager [req-402de077-dcd7-4ab9-b971-1ba51e5cca09 req-cfc51851-e0b1-41cf-83f8-02345f0c0588 service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Neutron deleted interface 96711ac0-bf70-468e-9ecc-70a5313e4bee; detaching it from the instance and deleting it from the info cache [ 841.277519] env[61906]: DEBUG nova.network.neutron [req-402de077-dcd7-4ab9-b971-1ba51e5cca09 req-cfc51851-e0b1-41cf-83f8-02345f0c0588 service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.284409] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 841.284409] env[61906]: value = "task-1356678" [ 841.284409] env[61906]: _type = "Task" [ 841.284409] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.289306] env[61906]: DEBUG nova.compute.manager [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 841.294241] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Acquiring lock "refresh_cache-0af2b07b-0632-4376-b502-58fe23d79f02" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.294849] env[61906]: DEBUG nova.network.neutron [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Successfully created port: ba127004-dc19-4fc4-b84b-97584ac68f34 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.303813] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356678, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.398901] env[61906]: DEBUG nova.network.neutron [req-41f117e3-c9a4-4d53-902c-05f0123ae196 req-b3294b28-7801-4de0-bd24-1b143d24748f service nova] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.525935] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356677, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.580445] env[61906]: DEBUG nova.network.neutron [req-41f117e3-c9a4-4d53-902c-05f0123ae196 req-b3294b28-7801-4de0-bd24-1b143d24748f service nova] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.607427] env[61906]: DEBUG nova.scheduler.client.report [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 841.721424] env[61906]: DEBUG nova.network.neutron [-] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.780802] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e990b10-1ee3-4bda-91d6-183a78e5d620 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.795776] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e0d2b3-3255-4ef7-80fb-5b8824965c43 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.815887] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356678, 'name': ReconfigVM_Task, 'duration_secs': 0.413031} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.816255] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44/31d0f640-47f4-4b72-b9fe-6ef4c76c6b44.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.816888] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5b4febe3-d040-40ec-9373-6d1349db917b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.824199] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 841.824199] env[61906]: value = "task-1356679" [ 841.824199] env[61906]: _type = "Task" [ 841.824199] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.832665] env[61906]: DEBUG nova.compute.manager [req-402de077-dcd7-4ab9-b971-1ba51e5cca09 req-cfc51851-e0b1-41cf-83f8-02345f0c0588 service nova] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Detach interface failed, port_id=96711ac0-bf70-468e-9ecc-70a5313e4bee, reason: Instance 0552e012-edcf-444c-a9b3-f1cea93b9e82 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 841.841311] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356679, 'name': Rename_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.026143] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356677, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.082894] env[61906]: DEBUG oslo_concurrency.lockutils [req-41f117e3-c9a4-4d53-902c-05f0123ae196 req-b3294b28-7801-4de0-bd24-1b143d24748f service nova] Releasing lock "refresh_cache-0af2b07b-0632-4376-b502-58fe23d79f02" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.082994] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Acquired lock "refresh_cache-0af2b07b-0632-4376-b502-58fe23d79f02" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.083158] env[61906]: DEBUG nova.network.neutron [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.112805] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.832s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.115933] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.485s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.116809] env[61906]: INFO nova.compute.claims [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.171630] env[61906]: INFO nova.scheduler.client.report [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleted allocations for instance 9a23071d-40fd-4446-aa03-ecbddcafe3d6 [ 842.224481] env[61906]: INFO nova.compute.manager [-] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Took 1.48 seconds to deallocate network for instance. [ 842.313089] env[61906]: DEBUG nova.compute.manager [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 842.343587] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356679, 'name': Rename_Task, 'duration_secs': 0.379826} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.344121] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.344425] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-344e196b-04e4-4bc7-8f6a-0c805fea1830 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.351643] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 842.351643] env[61906]: value = "task-1356680" [ 842.351643] env[61906]: _type = "Task" [ 842.351643] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.361260] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356680, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.363308] env[61906]: DEBUG nova.virt.hardware [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.363534] env[61906]: DEBUG nova.virt.hardware [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.363692] env[61906]: DEBUG nova.virt.hardware [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.363879] env[61906]: DEBUG nova.virt.hardware [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.364048] env[61906]: DEBUG nova.virt.hardware [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.364179] env[61906]: DEBUG nova.virt.hardware [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.364386] env[61906]: DEBUG nova.virt.hardware [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.364544] env[61906]: DEBUG nova.virt.hardware [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.364708] env[61906]: DEBUG nova.virt.hardware [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.364871] env[61906]: DEBUG nova.virt.hardware [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.365048] env[61906]: DEBUG nova.virt.hardware [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.365819] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f031b282-7b14-4d71-a792-bedf669ba4e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.372875] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a041d663-7b1b-476b-b509-f3d72434029f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.472459] env[61906]: INFO nova.compute.manager [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Unrescuing [ 842.472870] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.473091] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.473298] env[61906]: DEBUG nova.network.neutron [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.526990] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356677, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.637519] env[61906]: DEBUG nova.network.neutron [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.680107] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dd7ff0d-e420-4eda-af21-3bce89ce5142 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "9a23071d-40fd-4446-aa03-ecbddcafe3d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.849s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.732154] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.862387] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356680, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.887647] env[61906]: DEBUG nova.network.neutron [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Updating instance_info_cache with network_info: [{"id": "fcbc3598-49e0-4da8-91e1-e0f6929f305f", "address": "fa:16:3e:02:6b:d5", "network": {"id": "6ae5a369-7117-4fa3-bc2d-b5594c32b424", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-188819444-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9da9bb542c6a4aba9e99b402d7f89577", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcbc3598-49", "ovs_interfaceid": "fcbc3598-49e0-4da8-91e1-e0f6929f305f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.038185] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356677, 'name': CloneVM_Task, 'duration_secs': 1.913138} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.038185] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Created linked-clone VM from snapshot [ 843.038444] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cbe37e9-289e-471a-94fa-e81b6aef35ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.052584] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Uploading image df684168-2299-49f4-8c2b-403aedfc499a {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 843.084928] env[61906]: DEBUG oslo_vmware.rw_handles [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 843.084928] env[61906]: value = "vm-288982" [ 843.084928] env[61906]: _type = "VirtualMachine" [ 843.084928] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 843.085389] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-90cc3ddb-d09b-462e-b7a1-6c012a70c512 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.095095] env[61906]: DEBUG oslo_vmware.rw_handles [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lease: (returnval){ [ 843.095095] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52278fa4-0216-f571-6578-66fb0e3c97ec" [ 843.095095] env[61906]: _type = "HttpNfcLease" [ 843.095095] env[61906]: } obtained for exporting VM: (result){ [ 843.095095] env[61906]: value = "vm-288982" [ 843.095095] env[61906]: _type = "VirtualMachine" [ 843.095095] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 843.095411] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the lease: (returnval){ [ 843.095411] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52278fa4-0216-f571-6578-66fb0e3c97ec" [ 843.095411] env[61906]: _type = "HttpNfcLease" [ 843.095411] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 843.103672] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 843.103672] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52278fa4-0216-f571-6578-66fb0e3c97ec" [ 843.103672] env[61906]: _type = "HttpNfcLease" [ 843.103672] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 843.157328] env[61906]: DEBUG nova.network.neutron [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Successfully updated port: ba127004-dc19-4fc4-b84b-97584ac68f34 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.284143] env[61906]: DEBUG nova.compute.manager [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Received event network-vif-plugged-ba127004-dc19-4fc4-b84b-97584ac68f34 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.284143] env[61906]: DEBUG oslo_concurrency.lockutils [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] Acquiring lock "943c57f9-74c6-4465-9773-3fb01e78127b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.284143] env[61906]: DEBUG oslo_concurrency.lockutils [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] Lock "943c57f9-74c6-4465-9773-3fb01e78127b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.285296] env[61906]: DEBUG oslo_concurrency.lockutils [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] Lock "943c57f9-74c6-4465-9773-3fb01e78127b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.285647] env[61906]: DEBUG nova.compute.manager [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] No waiting events found dispatching network-vif-plugged-ba127004-dc19-4fc4-b84b-97584ac68f34 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 843.286088] env[61906]: WARNING nova.compute.manager [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Received unexpected event network-vif-plugged-ba127004-dc19-4fc4-b84b-97584ac68f34 for instance with vm_state building and task_state spawning. [ 843.286360] env[61906]: DEBUG nova.compute.manager [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Received event network-changed-ba127004-dc19-4fc4-b84b-97584ac68f34 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.286932] env[61906]: DEBUG nova.compute.manager [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Refreshing instance network info cache due to event network-changed-ba127004-dc19-4fc4-b84b-97584ac68f34. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 843.287250] env[61906]: DEBUG oslo_concurrency.lockutils [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] Acquiring lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.287509] env[61906]: DEBUG oslo_concurrency.lockutils [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] Acquired lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.288393] env[61906]: DEBUG nova.network.neutron [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Refreshing network info cache for port ba127004-dc19-4fc4-b84b-97584ac68f34 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 843.324887] env[61906]: DEBUG nova.network.neutron [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Updating instance_info_cache with network_info: [{"id": "e291dedf-5c3f-4d78-bad3-295b2d1e2f75", "address": "fa:16:3e:c8:8d:ef", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape291dedf-5c", "ovs_interfaceid": "e291dedf-5c3f-4d78-bad3-295b2d1e2f75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.364065] env[61906]: DEBUG oslo_vmware.api [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356680, 'name': PowerOnVM_Task, 'duration_secs': 0.676821} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.364339] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.364537] env[61906]: INFO nova.compute.manager [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Took 5.89 seconds to spawn the instance on the hypervisor. [ 843.364710] env[61906]: DEBUG nova.compute.manager [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.365502] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e1f2d30-846c-489a-b593-58ec9d3b7449 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.370935] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf27528-8bca-48cc-8e0f-7c66279f6df9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.382421] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa91e591-0cc6-4b86-8269-3078e1e0f49f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.413250] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Releasing lock "refresh_cache-0af2b07b-0632-4376-b502-58fe23d79f02" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.413557] env[61906]: DEBUG nova.compute.manager [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Instance network_info: |[{"id": "fcbc3598-49e0-4da8-91e1-e0f6929f305f", "address": "fa:16:3e:02:6b:d5", "network": {"id": "6ae5a369-7117-4fa3-bc2d-b5594c32b424", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-188819444-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9da9bb542c6a4aba9e99b402d7f89577", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcbc3598-49", "ovs_interfaceid": "fcbc3598-49e0-4da8-91e1-e0f6929f305f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 843.414439] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:6b:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fcbc3598-49e0-4da8-91e1-e0f6929f305f', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.421516] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Creating folder: Project (9da9bb542c6a4aba9e99b402d7f89577). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.422224] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12069c2b-a3a0-4f00-95f0-7d0cab7f6bac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.424761] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0e1f102f-d7c6-4419-98fe-64c589878953 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.431716] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b82a0b0a-5925-434b-a0a9-16c935f19a64 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.437021] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Created folder: Project (9da9bb542c6a4aba9e99b402d7f89577) in parent group-v288914. [ 843.437021] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Creating folder: Instances. Parent ref: group-v288983. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.437436] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0dfdbe72-9571-470c-8e0b-a6f981b070d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.447451] env[61906]: DEBUG nova.compute.provider_tree [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.457788] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Created folder: Instances in parent group-v288983. [ 843.457788] env[61906]: DEBUG oslo.service.loopingcall [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.457788] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 843.457788] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0aca66b-a31a-46da-b8c4-0d28bf06f177 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.477175] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.477175] env[61906]: value = "task-1356684" [ 843.477175] env[61906]: _type = "Task" [ 843.477175] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.485014] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356684, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.606089] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 843.606089] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52278fa4-0216-f571-6578-66fb0e3c97ec" [ 843.606089] env[61906]: _type = "HttpNfcLease" [ 843.606089] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 843.606413] env[61906]: DEBUG oslo_vmware.rw_handles [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 843.606413] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52278fa4-0216-f571-6578-66fb0e3c97ec" [ 843.606413] env[61906]: _type = "HttpNfcLease" [ 843.606413] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 843.607247] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0485314e-a791-465c-9847-a96fcae7f7d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.615199] env[61906]: DEBUG oslo_vmware.rw_handles [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527f8f9d-3618-d08f-9391-3a30962572f3/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 843.615390] env[61906]: DEBUG oslo_vmware.rw_handles [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527f8f9d-3618-d08f-9391-3a30962572f3/disk-0.vmdk for reading. {{(pid=61906) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 843.673412] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.723535] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-12bf7442-95ae-42c4-bc8c-373521503cd2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.831093] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.832073] env[61906]: DEBUG nova.objects.instance [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lazy-loading 'flavor' on Instance uuid 79452791-59cb-4722-bb4a-8e59d8c4e641 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.891624] env[61906]: INFO nova.compute.manager [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Took 31.43 seconds to build instance. [ 843.916717] env[61906]: DEBUG nova.network.neutron [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.950572] env[61906]: DEBUG nova.scheduler.client.report [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 843.993300] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356684, 'name': CreateVM_Task, 'duration_secs': 0.476939} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.993484] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.994244] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.994567] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.995089] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.995406] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1baf8d13-6b3b-42dd-9f6c-307c4cbff37e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.001640] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Waiting for the task: (returnval){ [ 844.001640] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522f8ea2-525c-d9c0-ed3e-8944f52391aa" [ 844.001640] env[61906]: _type = "Task" [ 844.001640] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.013235] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522f8ea2-525c-d9c0-ed3e-8944f52391aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.041668] env[61906]: DEBUG nova.network.neutron [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.341780] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02fc0d58-ba65-4269-8011-55bbf4263a1a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.365035] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 844.365638] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ac9aebd-b8ef-4b8f-a949-d9e609137e79 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.375540] env[61906]: DEBUG oslo_vmware.api [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 844.375540] env[61906]: value = "task-1356685" [ 844.375540] env[61906]: _type = "Task" [ 844.375540] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.385561] env[61906]: DEBUG oslo_vmware.api [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356685, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.394367] env[61906]: DEBUG oslo_concurrency.lockutils [None req-867deea6-4659-4b8c-92f1-c9c7d2f484d6 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "31d0f640-47f4-4b72-b9fe-6ef4c76c6b44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.397s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.457142] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.457691] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 844.461202] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.807s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.462501] env[61906]: INFO nova.compute.claims [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.516297] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522f8ea2-525c-d9c0-ed3e-8944f52391aa, 'name': SearchDatastore_Task, 'duration_secs': 0.013953} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.516651] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.516949] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.517320] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.517439] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.517624] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.519148] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a4b88d7-3075-46b5-81ec-99a9c102add0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.530712] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.531028] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.531827] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa462980-f42f-4ca7-b906-5bed9239d291 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.539511] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Waiting for the task: (returnval){ [ 844.539511] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522919c8-87b5-7eb8-148e-72c77f6c3cbe" [ 844.539511] env[61906]: _type = "Task" [ 844.539511] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.544422] env[61906]: DEBUG oslo_concurrency.lockutils [req-3ef37fc2-40ce-418e-a1e5-b31df186bd6a req-3e5b6cfb-3e69-4a11-9f39-b8f0066dec90 service nova] Releasing lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.549245] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.550013] env[61906]: DEBUG nova.network.neutron [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.550851] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522919c8-87b5-7eb8-148e-72c77f6c3cbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.877197] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "33e288d5-9065-4606-b17a-c4c90bcbd533" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.877719] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "33e288d5-9065-4606-b17a-c4c90bcbd533" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.899168] env[61906]: DEBUG oslo_vmware.api [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356685, 'name': PowerOffVM_Task, 'duration_secs': 0.314603} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.900019] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 844.910317] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Reconfiguring VM instance instance-00000044 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 844.911806] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-957a2faf-8c77-4984-bc14-8e6ce817f926 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.938651] env[61906]: DEBUG oslo_vmware.api [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 844.938651] env[61906]: value = "task-1356686" [ 844.938651] env[61906]: _type = "Task" [ 844.938651] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.949839] env[61906]: DEBUG oslo_vmware.api [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356686, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.971022] env[61906]: DEBUG nova.compute.utils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 844.973038] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 844.974132] env[61906]: DEBUG nova.network.neutron [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 845.050432] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522919c8-87b5-7eb8-148e-72c77f6c3cbe, 'name': SearchDatastore_Task, 'duration_secs': 0.013215} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.053905] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76b74ec8-75b8-4585-9961-43d679115674 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.061651] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Waiting for the task: (returnval){ [ 845.061651] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f543c3-1a6f-9bc6-0298-b25d52841612" [ 845.061651] env[61906]: _type = "Task" [ 845.061651] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.063121] env[61906]: DEBUG nova.policy [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747d463bd48b48f0a928417917774ebd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a8ab66571abf4eb7a6431e4aa9d9dd3e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 845.073561] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f543c3-1a6f-9bc6-0298-b25d52841612, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.105500] env[61906]: DEBUG nova.network.neutron [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 845.242665] env[61906]: INFO nova.compute.manager [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Rebuilding instance [ 845.302609] env[61906]: DEBUG nova.compute.manager [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 845.303499] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7361d546-9c5f-447b-8585-bac6b48117d6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.323575] env[61906]: DEBUG nova.network.neutron [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance_info_cache with network_info: [{"id": "ba127004-dc19-4fc4-b84b-97584ac68f34", "address": "fa:16:3e:d1:ad:74", "network": {"id": "9e0705ba-7b9f-46fb-8603-01598a888212", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-705593104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb85e88fd4f54e0db2ff131f81137f64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba127004-dc", "ovs_interfaceid": "ba127004-dc19-4fc4-b84b-97584ac68f34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.387102] env[61906]: DEBUG nova.compute.manager [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 845.451983] env[61906]: DEBUG oslo_vmware.api [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356686, 'name': ReconfigVM_Task, 'duration_secs': 0.425204} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.452290] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Reconfigured VM instance instance-00000044 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 845.452484] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 845.452742] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d8c7f3ae-8d4a-4a3b-b403-5a46786eaeca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.462256] env[61906]: DEBUG oslo_vmware.api [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 845.462256] env[61906]: value = "task-1356687" [ 845.462256] env[61906]: _type = "Task" [ 845.462256] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.471044] env[61906]: DEBUG oslo_vmware.api [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356687, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.476697] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 845.559309] env[61906]: DEBUG nova.network.neutron [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Successfully created port: ba141737-69f3-4c17-a4af-b5c5a4447d6b {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 845.577648] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f543c3-1a6f-9bc6-0298-b25d52841612, 'name': SearchDatastore_Task, 'duration_secs': 0.012528} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.578397] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.578908] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 0af2b07b-0632-4376-b502-58fe23d79f02/0af2b07b-0632-4376-b502-58fe23d79f02.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 845.579212] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-61471a8f-c967-4c6a-82e6-cffd14e7c3be {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.588932] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Waiting for the task: (returnval){ [ 845.588932] env[61906]: value = "task-1356688" [ 845.588932] env[61906]: _type = "Task" [ 845.588932] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.603385] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356688, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.738811] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d10d1cc7-ddcb-499d-9146-26ef8d4cbca4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.747477] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c783ce1f-d924-4b2f-850f-0c5e5adc73ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.779563] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e60285-f0b4-44b8-ba65-bac7dacdc57c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.788684] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6350fdaa-d41d-4fa7-a2fe-9235e2b7f7f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.804349] env[61906]: DEBUG nova.compute.provider_tree [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.815184] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.815809] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06368254-7bde-447d-a721-90569c5fbe1e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.824397] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 845.824397] env[61906]: value = "task-1356689" [ 845.824397] env[61906]: _type = "Task" [ 845.824397] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.828527] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.828905] env[61906]: DEBUG nova.compute.manager [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Instance network_info: |[{"id": "ba127004-dc19-4fc4-b84b-97584ac68f34", "address": "fa:16:3e:d1:ad:74", "network": {"id": "9e0705ba-7b9f-46fb-8603-01598a888212", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-705593104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb85e88fd4f54e0db2ff131f81137f64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba127004-dc", "ovs_interfaceid": "ba127004-dc19-4fc4-b84b-97584ac68f34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 845.829479] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:ad:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4c5eb94-841c-4713-985a-8fc4117fbaf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ba127004-dc19-4fc4-b84b-97584ac68f34', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.841027] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Creating folder: Project (eb85e88fd4f54e0db2ff131f81137f64). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 845.841027] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bb1eb43e-fea8-416f-8b59-7c9d32730800 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.844351] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356689, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.851459] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Created folder: Project (eb85e88fd4f54e0db2ff131f81137f64) in parent group-v288914. [ 845.851616] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Creating folder: Instances. Parent ref: group-v288986. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 845.851891] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3427be6b-4ce4-483d-8803-b04098d3867d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.864041] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Created folder: Instances in parent group-v288986. [ 845.864357] env[61906]: DEBUG oslo.service.loopingcall [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.864554] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 845.864784] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f10b4b48-cb34-4f6b-9b58-239da9eb5171 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.889233] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.889233] env[61906]: value = "task-1356692" [ 845.889233] env[61906]: _type = "Task" [ 845.889233] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.904496] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356692, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.909877] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.974512] env[61906]: DEBUG oslo_vmware.api [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356687, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.102611] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356688, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.308224] env[61906]: DEBUG nova.scheduler.client.report [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.337104] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356689, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.399633] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356692, 'name': CreateVM_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.474159] env[61906]: DEBUG oslo_vmware.api [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356687, 'name': PowerOnVM_Task, 'duration_secs': 0.565301} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.475034] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 846.475034] env[61906]: DEBUG nova.compute.manager [None req-3b5aa412-37d6-4579-bf05-544903d359e5 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.475545] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb463487-0917-4ac3-8b6d-758a127e87d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.486574] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 846.523301] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 846.523301] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 846.523301] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 846.523301] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 846.523593] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 846.523821] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 846.524073] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 846.524269] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 846.524450] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 846.524619] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 846.524825] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 846.525828] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0493ee67-3252-46bd-93ef-8c913ffcc311 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.536669] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83ee0c5-e31a-44ef-8062-dafaa0d95f98 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.601563] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356688, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637644} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.601925] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 0af2b07b-0632-4376-b502-58fe23d79f02/0af2b07b-0632-4376-b502-58fe23d79f02.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.602183] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.602518] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0337ed25-1dd1-45f4-a7d6-cd384066c331 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.610290] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Waiting for the task: (returnval){ [ 846.610290] env[61906]: value = "task-1356693" [ 846.610290] env[61906]: _type = "Task" [ 846.610290] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.618774] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356693, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.815166] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.815705] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 846.818447] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.449s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.843615] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356689, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.901896] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356692, 'name': CreateVM_Task, 'duration_secs': 0.757567} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.902160] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 846.902952] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.903150] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.903480] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 846.903769] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2016190e-8a6e-44f0-afc1-66b7d8776d9c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.910293] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 846.910293] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a7f7ba-cb0d-b73c-fe56-960fc6fa0ac7" [ 846.910293] env[61906]: _type = "Task" [ 846.910293] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.918941] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a7f7ba-cb0d-b73c-fe56-960fc6fa0ac7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.120808] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356693, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.322073] env[61906]: DEBUG nova.compute.utils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 847.331187] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 847.331864] env[61906]: DEBUG nova.network.neutron [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 847.344016] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356689, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.422540] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a7f7ba-cb0d-b73c-fe56-960fc6fa0ac7, 'name': SearchDatastore_Task, 'duration_secs': 0.010505} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.422884] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.423149] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 847.423397] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.423549] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.423757] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 847.424066] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f854fe6-643d-43bc-b7b5-1c68e6c4599c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.445614] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 847.445887] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 847.446715] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfa02c91-0ff6-4166-9772-8c3aa39885e9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.452850] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 847.452850] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5207bfd0-68fe-7aec-cb83-38ce715d74ce" [ 847.452850] env[61906]: _type = "Task" [ 847.452850] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.454358] env[61906]: DEBUG nova.policy [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747d463bd48b48f0a928417917774ebd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a8ab66571abf4eb7a6431e4aa9d9dd3e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 847.464457] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5207bfd0-68fe-7aec-cb83-38ce715d74ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.623278] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356693, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.798910] env[61906]: DEBUG nova.compute.manager [req-b6a3148c-c073-44a4-8757-1a93edb99378 req-8b5b95ee-5759-41b1-8025-5dca527c4890 service nova] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Received event network-vif-plugged-ba141737-69f3-4c17-a4af-b5c5a4447d6b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.799320] env[61906]: DEBUG oslo_concurrency.lockutils [req-b6a3148c-c073-44a4-8757-1a93edb99378 req-8b5b95ee-5759-41b1-8025-5dca527c4890 service nova] Acquiring lock "63cbf664-078b-40c5-92bb-7a5a8d8aea3b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.799416] env[61906]: DEBUG oslo_concurrency.lockutils [req-b6a3148c-c073-44a4-8757-1a93edb99378 req-8b5b95ee-5759-41b1-8025-5dca527c4890 service nova] Lock "63cbf664-078b-40c5-92bb-7a5a8d8aea3b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.799549] env[61906]: DEBUG oslo_concurrency.lockutils [req-b6a3148c-c073-44a4-8757-1a93edb99378 req-8b5b95ee-5759-41b1-8025-5dca527c4890 service nova] Lock "63cbf664-078b-40c5-92bb-7a5a8d8aea3b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.799715] env[61906]: DEBUG nova.compute.manager [req-b6a3148c-c073-44a4-8757-1a93edb99378 req-8b5b95ee-5759-41b1-8025-5dca527c4890 service nova] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] No waiting events found dispatching network-vif-plugged-ba141737-69f3-4c17-a4af-b5c5a4447d6b {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 847.800161] env[61906]: WARNING nova.compute.manager [req-b6a3148c-c073-44a4-8757-1a93edb99378 req-8b5b95ee-5759-41b1-8025-5dca527c4890 service nova] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Received unexpected event network-vif-plugged-ba141737-69f3-4c17-a4af-b5c5a4447d6b for instance with vm_state building and task_state spawning. [ 847.832194] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 847.850396] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356689, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.868021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 5adaa660-b736-4c11-9141-846cf475ccd5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.868021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.868021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 79452791-59cb-4722-bb4a-8e59d8c4e641 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.868021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 7bae658e-b050-4639-b34e-c2671ef5c773 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.868021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance e133605d-c630-4b9e-a314-bf496c853710 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.868021] env[61906]: WARNING nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 0552e012-edcf-444c-a9b3-f1cea93b9e82 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 847.868021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 24b4a747-67aa-4388-aed6-cb646cd55765 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.868021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance cb879dc0-3af7-4279-aa28-66e2b8b4286d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.868021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance a5f780a2-0cb1-4da8-8276-82b9653bf6a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.868021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.868021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 0af2b07b-0632-4376-b502-58fe23d79f02 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.868021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 943c57f9-74c6-4465-9773-3fb01e78127b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.868021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 63cbf664-078b-40c5-92bb-7a5a8d8aea3b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.868021] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 3a6da3bd-8e28-4c23-8660-f32f5c862d02 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.969080] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5207bfd0-68fe-7aec-cb83-38ce715d74ce, 'name': SearchDatastore_Task, 'duration_secs': 0.011002} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.970196] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e89c5751-7b14-49f0-a148-7522b67fb30e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.974551] env[61906]: DEBUG nova.network.neutron [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Successfully updated port: ba141737-69f3-4c17-a4af-b5c5a4447d6b {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 847.983307] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 847.983307] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]523f91ca-63f7-d571-aac5-8bbcc74ed6a2" [ 847.983307] env[61906]: _type = "Task" [ 847.983307] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.992743] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]523f91ca-63f7-d571-aac5-8bbcc74ed6a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.029129] env[61906]: DEBUG nova.network.neutron [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Successfully created port: 3d71c788-b4a5-4f7a-8184-d9bbe379bd70 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 848.129555] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356693, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.05668} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.130019] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.131966] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9316b407-5add-4b26-b42e-df595416e703 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.171080] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 0af2b07b-0632-4376-b502-58fe23d79f02/0af2b07b-0632-4376-b502-58fe23d79f02.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.171626] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-beeac2d2-2e0a-4a23-a135-b4c9a28c36c6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.204822] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Waiting for the task: (returnval){ [ 848.204822] env[61906]: value = "task-1356694" [ 848.204822] env[61906]: _type = "Task" [ 848.204822] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.221435] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356694, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.346375] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356689, 'name': PowerOffVM_Task, 'duration_secs': 2.258472} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.346753] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 848.347042] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 848.347840] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb62524b-6c27-4e3f-988c-e46af9c2c3f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.355340] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 848.355677] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b1a568f-73b3-4cc2-b9dc-f01d94b0936d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.371081] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 33e288d5-9065-4606-b17a-c4c90bcbd533 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 848.371694] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 848.371694] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3008MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 848.387116] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 848.387374] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 848.387572] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Deleting the datastore file [datastore1] 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 848.387847] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7fa27cac-72cd-43ea-b2b7-453f3d701f00 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.395618] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 848.395618] env[61906]: value = "task-1356696" [ 848.395618] env[61906]: _type = "Task" [ 848.395618] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.407967] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356696, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.479037] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "refresh_cache-63cbf664-078b-40c5-92bb-7a5a8d8aea3b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.479037] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired lock "refresh_cache-63cbf664-078b-40c5-92bb-7a5a8d8aea3b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.479241] env[61906]: DEBUG nova.network.neutron [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 848.496912] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]523f91ca-63f7-d571-aac5-8bbcc74ed6a2, 'name': SearchDatastore_Task, 'duration_secs': 0.010064} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.497184] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.497518] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 943c57f9-74c6-4465-9773-3fb01e78127b/943c57f9-74c6-4465-9773-3fb01e78127b.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 848.497839] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db5d22cc-f879-4bdc-9eb2-ff087bd4d984 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.506186] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 848.506186] env[61906]: value = "task-1356697" [ 848.506186] env[61906]: _type = "Task" [ 848.506186] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.519755] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356697, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.603470] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5a9f75-8a18-4dcb-87aa-6e3b4a5e21c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.614597] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3c625b-e432-468e-aa52-2e64f68de662 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.645692] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff512ea-208d-4677-b8c9-a7c9ec5f7b2a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.654241] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63725361-3863-4ef1-983f-c4e8b5f7cc7a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.670979] env[61906]: DEBUG nova.compute.provider_tree [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.719814] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356694, 'name': ReconfigVM_Task, 'duration_secs': 0.314203} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.720313] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 0af2b07b-0632-4376-b502-58fe23d79f02/0af2b07b-0632-4376-b502-58fe23d79f02.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.721256] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b24ee2c9-10fc-4aa8-a092-53d506db009c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.730189] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Waiting for the task: (returnval){ [ 848.730189] env[61906]: value = "task-1356698" [ 848.730189] env[61906]: _type = "Task" [ 848.730189] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.743722] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356698, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.848627] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 848.877424] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 848.877768] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 848.878033] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 848.878285] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 848.878481] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 848.878684] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 848.878914] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 848.879116] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 848.879302] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 848.879498] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 848.879696] env[61906]: DEBUG nova.virt.hardware [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 848.880694] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6df80f-f8fa-4124-9339-7bc40044c2fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.891697] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144c1c63-5e33-482b-a888-a49a4fa9896e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.918824] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356696, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.113445} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.919190] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.919463] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 848.919684] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 849.018226] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356697, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.019230] env[61906]: DEBUG nova.network.neutron [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.174328] env[61906]: DEBUG nova.scheduler.client.report [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.180155] env[61906]: DEBUG nova.network.neutron [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Updating instance_info_cache with network_info: [{"id": "ba141737-69f3-4c17-a4af-b5c5a4447d6b", "address": "fa:16:3e:b3:f6:3e", "network": {"id": "ef77b791-3dbc-4771-b384-b07b20ffd012", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1332279450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8ab66571abf4eb7a6431e4aa9d9dd3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba141737-69", "ovs_interfaceid": "ba141737-69f3-4c17-a4af-b5c5a4447d6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.243196] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356698, 'name': Rename_Task, 'duration_secs': 0.207347} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.243644] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 849.243994] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ebaccecb-7a9d-4a45-bb2e-2f6a2caf18c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.252914] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Waiting for the task: (returnval){ [ 849.252914] env[61906]: value = "task-1356699" [ 849.252914] env[61906]: _type = "Task" [ 849.252914] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.261859] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356699, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.517637] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356697, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.58002} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.518084] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 943c57f9-74c6-4465-9773-3fb01e78127b/943c57f9-74c6-4465-9773-3fb01e78127b.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 849.518215] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.518486] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f76457c2-0120-45e1-b772-ed5088570890 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.527811] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 849.527811] env[61906]: value = "task-1356700" [ 849.527811] env[61906]: _type = "Task" [ 849.527811] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.538208] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356700, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.629625] env[61906]: DEBUG nova.network.neutron [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Successfully updated port: 3d71c788-b4a5-4f7a-8184-d9bbe379bd70 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.682106] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 849.682354] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.864s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.682640] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.951s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.682832] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.686171] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.776s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.688216] env[61906]: INFO nova.compute.claims [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.692712] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Releasing lock "refresh_cache-63cbf664-078b-40c5-92bb-7a5a8d8aea3b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.693484] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Instance network_info: |[{"id": "ba141737-69f3-4c17-a4af-b5c5a4447d6b", "address": "fa:16:3e:b3:f6:3e", "network": {"id": "ef77b791-3dbc-4771-b384-b07b20ffd012", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1332279450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8ab66571abf4eb7a6431e4aa9d9dd3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba141737-69", "ovs_interfaceid": "ba141737-69f3-4c17-a4af-b5c5a4447d6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 849.694602] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:f6:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '098df9b7-d759-47f7-b756-334848cb423b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ba141737-69f3-4c17-a4af-b5c5a4447d6b', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.703440] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Creating folder: Project (a8ab66571abf4eb7a6431e4aa9d9dd3e). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.705098] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-79ba51e4-e70e-4e84-9cbb-5ee1f58c3119 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.711855] env[61906]: INFO nova.scheduler.client.report [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Deleted allocations for instance 0552e012-edcf-444c-a9b3-f1cea93b9e82 [ 849.731147] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Created folder: Project (a8ab66571abf4eb7a6431e4aa9d9dd3e) in parent group-v288914. [ 849.731514] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Creating folder: Instances. Parent ref: group-v288989. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.731865] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7d9cd9c-cb10-45e4-8ac3-c690318b3091 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.748826] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Created folder: Instances in parent group-v288989. [ 849.749349] env[61906]: DEBUG oslo.service.loopingcall [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.750166] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.750481] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3296cb95-15ca-4567-aac5-97b7967fa466 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.778782] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356699, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.780410] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.780410] env[61906]: value = "task-1356703" [ 849.780410] env[61906]: _type = "Task" [ 849.780410] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.791209] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356703, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.853423] env[61906]: DEBUG nova.compute.manager [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Received event network-changed-ba141737-69f3-4c17-a4af-b5c5a4447d6b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.853892] env[61906]: DEBUG nova.compute.manager [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Refreshing instance network info cache due to event network-changed-ba141737-69f3-4c17-a4af-b5c5a4447d6b. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 849.855080] env[61906]: DEBUG oslo_concurrency.lockutils [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] Acquiring lock "refresh_cache-63cbf664-078b-40c5-92bb-7a5a8d8aea3b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.855383] env[61906]: DEBUG oslo_concurrency.lockutils [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] Acquired lock "refresh_cache-63cbf664-078b-40c5-92bb-7a5a8d8aea3b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.855657] env[61906]: DEBUG nova.network.neutron [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Refreshing network info cache for port ba141737-69f3-4c17-a4af-b5c5a4447d6b {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 849.958010] env[61906]: DEBUG nova.virt.hardware [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 849.958361] env[61906]: DEBUG nova.virt.hardware [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 849.958565] env[61906]: DEBUG nova.virt.hardware [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.958785] env[61906]: DEBUG nova.virt.hardware [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 849.958960] env[61906]: DEBUG nova.virt.hardware [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.959158] env[61906]: DEBUG nova.virt.hardware [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 849.959433] env[61906]: DEBUG nova.virt.hardware [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 849.959636] env[61906]: DEBUG nova.virt.hardware [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 849.959849] env[61906]: DEBUG nova.virt.hardware [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 849.960064] env[61906]: DEBUG nova.virt.hardware [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 849.960336] env[61906]: DEBUG nova.virt.hardware [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.961223] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a61dce8f-8d18-4557-ac38-55284518a10c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.971080] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7adb5647-3d7f-412d-8c79-6d01e82ed5cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.987040] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.992800] env[61906]: DEBUG oslo.service.loopingcall [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.993681] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.993993] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fda6bb72-13b1-4662-a11e-ad0d46a88166 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.012327] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.012327] env[61906]: value = "task-1356704" [ 850.012327] env[61906]: _type = "Task" [ 850.012327] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.022635] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356704, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.041267] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356700, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0734} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.041883] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 850.042560] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cf8301-4d47-461f-bf94-b2ea1b39019e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.068715] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 943c57f9-74c6-4465-9773-3fb01e78127b/943c57f9-74c6-4465-9773-3fb01e78127b.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.069208] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1b032bb-1b59-4764-8ede-837060311bd8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.093309] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 850.093309] env[61906]: value = "task-1356705" [ 850.093309] env[61906]: _type = "Task" [ 850.093309] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.106928] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356705, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.132145] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "refresh_cache-3a6da3bd-8e28-4c23-8660-f32f5c862d02" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.132381] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired lock "refresh_cache-3a6da3bd-8e28-4c23-8660-f32f5c862d02" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.132465] env[61906]: DEBUG nova.network.neutron [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 850.214369] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "interface-5adaa660-b736-4c11-9141-846cf475ccd5-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.214734] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-5adaa660-b736-4c11-9141-846cf475ccd5-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.215137] env[61906]: DEBUG nova.objects.instance [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'flavor' on Instance uuid 5adaa660-b736-4c11-9141-846cf475ccd5 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.221151] env[61906]: DEBUG oslo_concurrency.lockutils [None req-cf00cebd-bd7a-4b29-aa42-2fe6bfea1d32 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479 tempest-FloatingIPsAssociationNegativeTestJSON-1936723479-project-member] Lock "0552e012-edcf-444c-a9b3-f1cea93b9e82" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.102s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.279881] env[61906]: DEBUG oslo_vmware.api [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356699, 'name': PowerOnVM_Task, 'duration_secs': 0.690178} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.280178] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 850.280391] env[61906]: INFO nova.compute.manager [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Took 10.34 seconds to spawn the instance on the hypervisor. [ 850.280573] env[61906]: DEBUG nova.compute.manager [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 850.281403] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa4441e-9638-4432-9eff-73feb1409e3a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.292648] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356703, 'name': CreateVM_Task, 'duration_secs': 0.432568} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.295186] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 850.297815] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.298091] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.298525] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.299039] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cb2e3a5-3063-4a15-aaea-bb5214c9fc2d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.305069] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 850.305069] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5217dd01-33b6-c642-03b7-95bc6afc407c" [ 850.305069] env[61906]: _type = "Task" [ 850.305069] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.314447] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5217dd01-33b6-c642-03b7-95bc6afc407c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.523039] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356704, 'name': CreateVM_Task, 'duration_secs': 0.351419} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.523254] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 850.523802] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.605020] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356705, 'name': ReconfigVM_Task, 'duration_secs': 0.383008} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.605555] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 943c57f9-74c6-4465-9773-3fb01e78127b/943c57f9-74c6-4465-9773-3fb01e78127b.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.606336] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f950ecb-4c64-4f5c-9c5c-a42751e5f244 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.617130] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 850.617130] env[61906]: value = "task-1356706" [ 850.617130] env[61906]: _type = "Task" [ 850.617130] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.627601] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356706, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.710853] env[61906]: DEBUG nova.network.neutron [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.816646] env[61906]: DEBUG nova.objects.instance [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'pci_requests' on Instance uuid 5adaa660-b736-4c11-9141-846cf475ccd5 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.818143] env[61906]: INFO nova.compute.manager [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Took 38.14 seconds to build instance. [ 850.831240] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5217dd01-33b6-c642-03b7-95bc6afc407c, 'name': SearchDatastore_Task, 'duration_secs': 0.011138} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.832470] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.833213] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.833213] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.833372] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.833672] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.834089] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.834547] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.838018] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b68fc631-87a8-40a8-8917-e01bd6297f59 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.841071] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1dd90a9-2d48-42e9-bc66-decdf053da1c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.849815] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 850.849815] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52057398-d325-6a61-8613-d7fb538b5188" [ 850.849815] env[61906]: _type = "Task" [ 850.849815] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.863262] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.863730] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.866727] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c89c207-cbb1-4427-be1b-3937a80167dc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.882134] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52057398-d325-6a61-8613-d7fb538b5188, 'name': SearchDatastore_Task, 'duration_secs': 0.016129} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.883226] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.884268] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.884652] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.891101] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 850.891101] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5254829e-04d3-b433-e5bf-3b6361654006" [ 850.891101] env[61906]: _type = "Task" [ 850.891101] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.904603] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5254829e-04d3-b433-e5bf-3b6361654006, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.937651] env[61906]: DEBUG nova.network.neutron [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Updated VIF entry in instance network info cache for port ba141737-69f3-4c17-a4af-b5c5a4447d6b. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 850.938390] env[61906]: DEBUG nova.network.neutron [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Updating instance_info_cache with network_info: [{"id": "ba141737-69f3-4c17-a4af-b5c5a4447d6b", "address": "fa:16:3e:b3:f6:3e", "network": {"id": "ef77b791-3dbc-4771-b384-b07b20ffd012", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1332279450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8ab66571abf4eb7a6431e4aa9d9dd3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba141737-69", "ovs_interfaceid": "ba141737-69f3-4c17-a4af-b5c5a4447d6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.974463] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f801eb-e546-4e78-9a4a-955c778865e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.984131] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-787fcb1d-6bda-4b2e-abc5-422d6d921a08 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.016026] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa77bf5f-e66f-4aa2-b29e-9cd6ad0725bb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.025107] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3ac88b-4455-45d3-a4d2-e415ee78beb8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.039645] env[61906]: DEBUG nova.compute.provider_tree [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.076119] env[61906]: DEBUG nova.network.neutron [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Updating instance_info_cache with network_info: [{"id": "3d71c788-b4a5-4f7a-8184-d9bbe379bd70", "address": "fa:16:3e:91:16:3c", "network": {"id": "ef77b791-3dbc-4771-b384-b07b20ffd012", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1332279450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8ab66571abf4eb7a6431e4aa9d9dd3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d71c788-b4", "ovs_interfaceid": "3d71c788-b4a5-4f7a-8184-d9bbe379bd70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.128160] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356706, 'name': Rename_Task, 'duration_secs': 0.179055} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.128490] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 851.128748] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c3383369-af78-40ff-bd6c-b7d3126e2813 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.137637] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 851.137637] env[61906]: value = "task-1356707" [ 851.137637] env[61906]: _type = "Task" [ 851.137637] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.148860] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356707, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.320777] env[61906]: DEBUG nova.objects.base [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Object Instance<5adaa660-b736-4c11-9141-846cf475ccd5> lazy-loaded attributes: flavor,pci_requests {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 851.321127] env[61906]: DEBUG nova.network.neutron [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 851.323211] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97b703a3-814e-4ff4-929a-1f5449964a24 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Lock "0af2b07b-0632-4376-b502-58fe23d79f02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.972s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.404084] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5254829e-04d3-b433-e5bf-3b6361654006, 'name': SearchDatastore_Task, 'duration_secs': 0.028784} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.405968] env[61906]: DEBUG nova.policy [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73219f5e96f64d1da86a19da9b18b06a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5533be74874b4094b7ef0f98a121f60b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 851.408954] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c499fd5c-70a7-42ee-aff9-6637e2f57a2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.420221] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.420575] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.421847] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 851.421847] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52398041-1487-6ebf-286e-daac1dd34dd1" [ 851.421847] env[61906]: _type = "Task" [ 851.421847] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.432648] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52398041-1487-6ebf-286e-daac1dd34dd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.441144] env[61906]: DEBUG oslo_concurrency.lockutils [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] Releasing lock "refresh_cache-63cbf664-078b-40c5-92bb-7a5a8d8aea3b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.441360] env[61906]: DEBUG nova.compute.manager [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Received event network-vif-plugged-3d71c788-b4a5-4f7a-8184-d9bbe379bd70 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.441595] env[61906]: DEBUG oslo_concurrency.lockutils [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] Acquiring lock "3a6da3bd-8e28-4c23-8660-f32f5c862d02-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.441812] env[61906]: DEBUG oslo_concurrency.lockutils [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] Lock "3a6da3bd-8e28-4c23-8660-f32f5c862d02-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.442023] env[61906]: DEBUG oslo_concurrency.lockutils [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] Lock "3a6da3bd-8e28-4c23-8660-f32f5c862d02-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.442261] env[61906]: DEBUG nova.compute.manager [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] No waiting events found dispatching network-vif-plugged-3d71c788-b4a5-4f7a-8184-d9bbe379bd70 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 851.442570] env[61906]: WARNING nova.compute.manager [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Received unexpected event network-vif-plugged-3d71c788-b4a5-4f7a-8184-d9bbe379bd70 for instance with vm_state building and task_state spawning. [ 851.442668] env[61906]: DEBUG nova.compute.manager [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Received event network-changed-3d71c788-b4a5-4f7a-8184-d9bbe379bd70 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.442864] env[61906]: DEBUG nova.compute.manager [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Refreshing instance network info cache due to event network-changed-3d71c788-b4a5-4f7a-8184-d9bbe379bd70. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 851.443071] env[61906]: DEBUG oslo_concurrency.lockutils [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] Acquiring lock "refresh_cache-3a6da3bd-8e28-4c23-8660-f32f5c862d02" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.542659] env[61906]: DEBUG nova.scheduler.client.report [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.579374] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Releasing lock "refresh_cache-3a6da3bd-8e28-4c23-8660-f32f5c862d02" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.579711] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Instance network_info: |[{"id": "3d71c788-b4a5-4f7a-8184-d9bbe379bd70", "address": "fa:16:3e:91:16:3c", "network": {"id": "ef77b791-3dbc-4771-b384-b07b20ffd012", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1332279450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8ab66571abf4eb7a6431e4aa9d9dd3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d71c788-b4", "ovs_interfaceid": "3d71c788-b4a5-4f7a-8184-d9bbe379bd70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 851.580348] env[61906]: DEBUG oslo_concurrency.lockutils [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] Acquired lock "refresh_cache-3a6da3bd-8e28-4c23-8660-f32f5c862d02" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.580348] env[61906]: DEBUG nova.network.neutron [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Refreshing network info cache for port 3d71c788-b4a5-4f7a-8184-d9bbe379bd70 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 851.581329] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:16:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '098df9b7-d759-47f7-b756-334848cb423b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d71c788-b4a5-4f7a-8184-d9bbe379bd70', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.589128] env[61906]: DEBUG oslo.service.loopingcall [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.589647] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 851.589836] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dcb7b76b-8383-4d20-b2c8-6bd8e2d700d9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.612068] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.612068] env[61906]: value = "task-1356708" [ 851.612068] env[61906]: _type = "Task" [ 851.612068] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.624350] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356708, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.648895] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356707, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.901111] env[61906]: DEBUG nova.network.neutron [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Successfully created port: 38287aaf-c4d5-440a-b756-6502a2c7dc02 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 851.923473] env[61906]: DEBUG nova.compute.manager [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 851.941793] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52398041-1487-6ebf-286e-daac1dd34dd1, 'name': SearchDatastore_Task, 'duration_secs': 0.016002} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.941979] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.942156] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 63cbf664-078b-40c5-92bb-7a5a8d8aea3b/63cbf664-078b-40c5-92bb-7a5a8d8aea3b.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.942459] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.942650] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.943433] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1a8b2be6-5ddf-449e-a4d1-a15081e03ecf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.945338] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62561c45-3ab8-4613-b97b-4b8dca97a6c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.955269] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 851.955269] env[61906]: value = "task-1356709" [ 851.955269] env[61906]: _type = "Task" [ 851.955269] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.956844] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.957043] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 851.961381] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f09eb9b-eae5-4cf0-8204-8f30457a46a9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.971100] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356709, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.972270] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 851.972270] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a7cac7-d6cc-ac7e-f9ee-d94bb9651681" [ 851.972270] env[61906]: _type = "Task" [ 851.972270] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.984400] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a7cac7-d6cc-ac7e-f9ee-d94bb9651681, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.049022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.362s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.049022] env[61906]: DEBUG nova.compute.manager [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 852.125428] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356708, 'name': CreateVM_Task, 'duration_secs': 0.467794} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.125756] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 852.127071] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.127330] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.127975] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 852.128482] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9de5acb0-920a-491b-bdd4-e83ae3b987e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.137540] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 852.137540] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529ebc73-afd7-9079-2778-33ea7031989b" [ 852.137540] env[61906]: _type = "Task" [ 852.137540] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.158109] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529ebc73-afd7-9079-2778-33ea7031989b, 'name': SearchDatastore_Task, 'duration_secs': 0.018082} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.161606] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.161949] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.162296] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.162801] env[61906]: DEBUG oslo_vmware.api [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356707, 'name': PowerOnVM_Task, 'duration_secs': 0.740996} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.163216] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 852.163286] env[61906]: INFO nova.compute.manager [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Took 9.85 seconds to spawn the instance on the hypervisor. [ 852.163528] env[61906]: DEBUG nova.compute.manager [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.164854] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb75344-ad28-49fc-90de-891e9a73c0dd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.464677] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.465109] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.467377] env[61906]: INFO nova.compute.claims [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.486433] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356709, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.495893] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a7cac7-d6cc-ac7e-f9ee-d94bb9651681, 'name': SearchDatastore_Task, 'duration_secs': 0.021899} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.496259] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab6701de-0370-4f1a-bdbc-7ef163000aed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.507411] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 852.507411] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52242ad3-484b-690b-421b-ce7813a42f06" [ 852.507411] env[61906]: _type = "Task" [ 852.507411] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.521049] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52242ad3-484b-690b-421b-ce7813a42f06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.555992] env[61906]: DEBUG nova.compute.utils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 852.559444] env[61906]: DEBUG nova.compute.manager [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 852.559954] env[61906]: DEBUG nova.network.neutron [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 852.563940] env[61906]: DEBUG nova.network.neutron [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Updated VIF entry in instance network info cache for port 3d71c788-b4a5-4f7a-8184-d9bbe379bd70. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 852.564413] env[61906]: DEBUG nova.network.neutron [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Updating instance_info_cache with network_info: [{"id": "3d71c788-b4a5-4f7a-8184-d9bbe379bd70", "address": "fa:16:3e:91:16:3c", "network": {"id": "ef77b791-3dbc-4771-b384-b07b20ffd012", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1332279450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8ab66571abf4eb7a6431e4aa9d9dd3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d71c788-b4", "ovs_interfaceid": "3d71c788-b4a5-4f7a-8184-d9bbe379bd70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.622438] env[61906]: DEBUG nova.policy [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf4fea47b929461785f4a487a56977a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3093466de87a45a2ac673c25bea19bc3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 852.627591] env[61906]: DEBUG oslo_vmware.rw_handles [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527f8f9d-3618-d08f-9391-3a30962572f3/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 852.629035] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6680fa03-34af-43d1-8dba-97ecc688e359 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.637186] env[61906]: DEBUG oslo_vmware.rw_handles [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527f8f9d-3618-d08f-9391-3a30962572f3/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 852.637417] env[61906]: ERROR oslo_vmware.rw_handles [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527f8f9d-3618-d08f-9391-3a30962572f3/disk-0.vmdk due to incomplete transfer. [ 852.637638] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-eb420311-645b-4662-9979-eed4878c6a61 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.646658] env[61906]: DEBUG oslo_vmware.rw_handles [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527f8f9d-3618-d08f-9391-3a30962572f3/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 852.646844] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Uploaded image df684168-2299-49f4-8c2b-403aedfc499a to the Glance image server {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 852.649435] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 852.649728] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b17b6684-b37d-427f-972c-fdaf042f7f3c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.658953] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 852.658953] env[61906]: value = "task-1356710" [ 852.658953] env[61906]: _type = "Task" [ 852.658953] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.669382] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356710, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.683321] env[61906]: INFO nova.compute.manager [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Took 33.29 seconds to build instance. [ 852.910943] env[61906]: DEBUG nova.network.neutron [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Successfully created port: a88ac5a5-d8da-4ebb-a745-ecc89275cb78 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.969959] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356709, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.719618} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.970066] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 63cbf664-078b-40c5-92bb-7a5a8d8aea3b/63cbf664-078b-40c5-92bb-7a5a8d8aea3b.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 852.970265] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.970531] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86deaf82-5d3a-42ef-9a4d-11733ea19627 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.983550] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 852.983550] env[61906]: value = "task-1356711" [ 852.983550] env[61906]: _type = "Task" [ 852.983550] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.993852] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356711, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.027346] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52242ad3-484b-690b-421b-ce7813a42f06, 'name': SearchDatastore_Task, 'duration_secs': 0.058374} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.027725] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.028117] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44/31d0f640-47f4-4b72-b9fe-6ef4c76c6b44.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 853.028453] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.028710] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.028911] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc73ad21-39d3-4dc0-bc0b-c2adbfa48168 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.032443] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6bc6494f-62dc-4488-89a6-8483fbfce578 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.041721] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 853.041721] env[61906]: value = "task-1356712" [ 853.041721] env[61906]: _type = "Task" [ 853.041721] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.047909] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.048157] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 853.049344] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-489b11da-d390-4edc-80bb-8ad17887eaed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.058368] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356712, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.063227] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 853.063227] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52db9618-bc53-6506-1ffb-95f6140025f5" [ 853.063227] env[61906]: _type = "Task" [ 853.063227] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.069054] env[61906]: DEBUG nova.compute.manager [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 853.072273] env[61906]: DEBUG oslo_concurrency.lockutils [req-e1052e4a-8acc-40dc-a1db-347ef641aa25 req-356fc28c-e415-45e8-9b3d-e5497ca77b52 service nova] Releasing lock "refresh_cache-3a6da3bd-8e28-4c23-8660-f32f5c862d02" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.078763] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52db9618-bc53-6506-1ffb-95f6140025f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.170321] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356710, 'name': Destroy_Task, 'duration_secs': 0.419451} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.170481] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Destroyed the VM [ 853.170719] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 853.171030] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3f40ff17-9ac0-4f5d-8316-f9534ec5277f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.176531] env[61906]: DEBUG oslo_concurrency.lockutils [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Acquiring lock "0af2b07b-0632-4376-b502-58fe23d79f02" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.176531] env[61906]: DEBUG oslo_concurrency.lockutils [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Lock "0af2b07b-0632-4376-b502-58fe23d79f02" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.176531] env[61906]: DEBUG oslo_concurrency.lockutils [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Acquiring lock "0af2b07b-0632-4376-b502-58fe23d79f02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.176531] env[61906]: DEBUG oslo_concurrency.lockutils [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Lock "0af2b07b-0632-4376-b502-58fe23d79f02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.176815] env[61906]: DEBUG oslo_concurrency.lockutils [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Lock "0af2b07b-0632-4376-b502-58fe23d79f02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.181047] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 853.181047] env[61906]: value = "task-1356713" [ 853.181047] env[61906]: _type = "Task" [ 853.181047] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.186590] env[61906]: INFO nova.compute.manager [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Terminating instance [ 853.189954] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1e3e375-5b48-431c-b9bb-2a8e8d048f01 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "943c57f9-74c6-4465-9773-3fb01e78127b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.870s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.190220] env[61906]: DEBUG nova.compute.manager [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 853.190314] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 853.192241] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea8f85ee-0c81-45db-9051-f883d12fabe7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.198218] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356713, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.203764] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.204072] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c5dfca1-530f-47b6-ab67-73ea6cc5b2b7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.211774] env[61906]: DEBUG oslo_vmware.api [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Waiting for the task: (returnval){ [ 853.211774] env[61906]: value = "task-1356714" [ 853.211774] env[61906]: _type = "Task" [ 853.211774] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.221201] env[61906]: DEBUG oslo_vmware.api [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356714, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.494394] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356711, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071426} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.494745] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 853.495668] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28af1f3a-857f-4adb-90ee-200641db2a42 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.523639] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 63cbf664-078b-40c5-92bb-7a5a8d8aea3b/63cbf664-078b-40c5-92bb-7a5a8d8aea3b.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 853.526857] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0db7657b-2bc1-40f0-8b5d-e87fd36934ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.550096] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 853.550096] env[61906]: value = "task-1356715" [ 853.550096] env[61906]: _type = "Task" [ 853.550096] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.556074] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356712, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.583932] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356715, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.595307] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52db9618-bc53-6506-1ffb-95f6140025f5, 'name': SearchDatastore_Task, 'duration_secs': 0.016033} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.597457] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d0a1850-9548-49f0-804e-370022d7d653 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.607149] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 853.607149] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525ef5de-9c85-d5bf-52a6-6b907141fd48" [ 853.607149] env[61906]: _type = "Task" [ 853.607149] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.624194] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525ef5de-9c85-d5bf-52a6-6b907141fd48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.691573] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356713, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.724485] env[61906]: DEBUG oslo_vmware.api [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356714, 'name': PowerOffVM_Task, 'duration_secs': 0.244878} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.724485] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 853.724485] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 853.724783] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-196b801a-a9fa-46fb-8cc8-dac790998d01 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.808228] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c0b895a-d408-418f-9087-2ae284be41b1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.817966] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 853.818160] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 853.818710] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Deleting the datastore file [datastore1] 0af2b07b-0632-4376-b502-58fe23d79f02 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 853.821168] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95af389f-4d98-4934-ae61-d3c1ea0464bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.826415] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f6e12a-e934-46e1-83cc-fb418734f88f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.835983] env[61906]: DEBUG oslo_vmware.api [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Waiting for the task: (returnval){ [ 853.835983] env[61906]: value = "task-1356717" [ 853.835983] env[61906]: _type = "Task" [ 853.835983] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.884747] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ef8c01-b2a8-4b0b-a424-624921938be6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.897364] env[61906]: DEBUG oslo_vmware.api [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356717, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.902933] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105a5556-d51c-46c5-98a8-432594f8177c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.929472] env[61906]: DEBUG nova.compute.provider_tree [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.949034] env[61906]: DEBUG nova.network.neutron [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Successfully updated port: 38287aaf-c4d5-440a-b756-6502a2c7dc02 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 853.990123] env[61906]: DEBUG nova.compute.manager [req-966128db-6de2-4323-bddf-deda6788ad8f req-0ae5163d-912c-4c6a-a100-8a42ed8f7d11 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Received event network-vif-plugged-38287aaf-c4d5-440a-b756-6502a2c7dc02 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.990422] env[61906]: DEBUG oslo_concurrency.lockutils [req-966128db-6de2-4323-bddf-deda6788ad8f req-0ae5163d-912c-4c6a-a100-8a42ed8f7d11 service nova] Acquiring lock "5adaa660-b736-4c11-9141-846cf475ccd5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.990673] env[61906]: DEBUG oslo_concurrency.lockutils [req-966128db-6de2-4323-bddf-deda6788ad8f req-0ae5163d-912c-4c6a-a100-8a42ed8f7d11 service nova] Lock "5adaa660-b736-4c11-9141-846cf475ccd5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.990894] env[61906]: DEBUG oslo_concurrency.lockutils [req-966128db-6de2-4323-bddf-deda6788ad8f req-0ae5163d-912c-4c6a-a100-8a42ed8f7d11 service nova] Lock "5adaa660-b736-4c11-9141-846cf475ccd5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.991141] env[61906]: DEBUG nova.compute.manager [req-966128db-6de2-4323-bddf-deda6788ad8f req-0ae5163d-912c-4c6a-a100-8a42ed8f7d11 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] No waiting events found dispatching network-vif-plugged-38287aaf-c4d5-440a-b756-6502a2c7dc02 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 853.991359] env[61906]: WARNING nova.compute.manager [req-966128db-6de2-4323-bddf-deda6788ad8f req-0ae5163d-912c-4c6a-a100-8a42ed8f7d11 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Received unexpected event network-vif-plugged-38287aaf-c4d5-440a-b756-6502a2c7dc02 for instance with vm_state active and task_state None. [ 854.055058] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356712, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.884451} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.055487] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44/31d0f640-47f4-4b72-b9fe-6ef4c76c6b44.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.055894] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.057098] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f4cbcf95-7ea3-4322-bc8d-f3ed57ee4ca2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.071584] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356715, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.073655] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 854.073655] env[61906]: value = "task-1356718" [ 854.073655] env[61906]: _type = "Task" [ 854.073655] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.085493] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356718, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.097317] env[61906]: DEBUG nova.compute.manager [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 854.124262] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525ef5de-9c85-d5bf-52a6-6b907141fd48, 'name': SearchDatastore_Task, 'duration_secs': 0.05192} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.126619] env[61906]: DEBUG nova.virt.hardware [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.126859] env[61906]: DEBUG nova.virt.hardware [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.127039] env[61906]: DEBUG nova.virt.hardware [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.127239] env[61906]: DEBUG nova.virt.hardware [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.127396] env[61906]: DEBUG nova.virt.hardware [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.127549] env[61906]: DEBUG nova.virt.hardware [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.127763] env[61906]: DEBUG nova.virt.hardware [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.128283] env[61906]: DEBUG nova.virt.hardware [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.128283] env[61906]: DEBUG nova.virt.hardware [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.128406] env[61906]: DEBUG nova.virt.hardware [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.128488] env[61906]: DEBUG nova.virt.hardware [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.128784] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.129042] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 3a6da3bd-8e28-4c23-8660-f32f5c862d02/3a6da3bd-8e28-4c23-8660-f32f5c862d02.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 854.129821] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee4c6e6-cfe7-45c4-843f-06bebd636ede {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.132506] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da8cacd7-7eba-4d80-bcd1-7bdf374fabf0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.142605] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8856280-f9a1-4260-8bae-750f74289389 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.148254] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 854.148254] env[61906]: value = "task-1356719" [ 854.148254] env[61906]: _type = "Task" [ 854.148254] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.173890] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356719, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.191057] env[61906]: DEBUG oslo_vmware.api [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356713, 'name': RemoveSnapshot_Task, 'duration_secs': 0.848711} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.191365] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 854.191611] env[61906]: INFO nova.compute.manager [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Took 15.79 seconds to snapshot the instance on the hypervisor. [ 854.243633] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "24b4a747-67aa-4388-aed6-cb646cd55765" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.244121] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "24b4a747-67aa-4388-aed6-cb646cd55765" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.244482] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "24b4a747-67aa-4388-aed6-cb646cd55765-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.244812] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "24b4a747-67aa-4388-aed6-cb646cd55765-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.245131] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "24b4a747-67aa-4388-aed6-cb646cd55765-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.253288] env[61906]: INFO nova.compute.manager [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Terminating instance [ 854.255527] env[61906]: DEBUG nova.compute.manager [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 854.256318] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 854.257969] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5886e4-de84-48c6-be33-7c7d5ef36504 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.274122] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 854.274649] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a41f25a6-ace0-411b-b123-9482ec05b11a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.378855] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 854.379264] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 854.379504] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleting the datastore file [datastore2] 24b4a747-67aa-4388-aed6-cb646cd55765 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.385976] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-299f8750-3378-420c-8a76-0e5f759c2e45 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.397216] env[61906]: DEBUG oslo_vmware.api [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Task: {'id': task-1356717, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.505921} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.398767] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.398971] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 854.399195] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 854.399385] env[61906]: INFO nova.compute.manager [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Took 1.21 seconds to destroy the instance on the hypervisor. [ 854.399671] env[61906]: DEBUG oslo.service.loopingcall [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.399983] env[61906]: DEBUG oslo_vmware.api [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 854.399983] env[61906]: value = "task-1356721" [ 854.399983] env[61906]: _type = "Task" [ 854.399983] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.400227] env[61906]: DEBUG nova.compute.manager [-] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 854.400325] env[61906]: DEBUG nova.network.neutron [-] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 854.411563] env[61906]: DEBUG oslo_vmware.api [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356721, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.433787] env[61906]: DEBUG nova.scheduler.client.report [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.454495] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.454700] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.454921] env[61906]: DEBUG nova.network.neutron [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 854.570654] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356715, 'name': ReconfigVM_Task, 'duration_secs': 0.950575} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.571220] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 63cbf664-078b-40c5-92bb-7a5a8d8aea3b/63cbf664-078b-40c5-92bb-7a5a8d8aea3b.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.571994] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e7155e8-e66d-44d2-b0c6-061635b06999 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.587321] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356718, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.188786} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.588927] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.589311] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 854.589311] env[61906]: value = "task-1356722" [ 854.589311] env[61906]: _type = "Task" [ 854.589311] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.590881] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aea31ac-20c4-459d-910b-671d60aa7684 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.605904] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356722, 'name': Rename_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.623686] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44/31d0f640-47f4-4b72-b9fe-6ef4c76c6b44.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.624071] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ddfca7f9-e090-4da2-a258-89abc498fc88 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.650968] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 854.650968] env[61906]: value = "task-1356723" [ 854.650968] env[61906]: _type = "Task" [ 854.650968] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.674495] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356723, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.674776] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356719, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.675361] env[61906]: DEBUG nova.compute.manager [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Stashing vm_state: active {{(pid=61906) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 854.698061] env[61906]: DEBUG nova.compute.manager [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Instance disappeared during snapshot {{(pid=61906) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 854.714048] env[61906]: DEBUG nova.compute.manager [None req-b742448d-f238-4ed3-b244-d5dd756f96f6 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Image not found during clean up df684168-2299-49f4-8c2b-403aedfc499a {{(pid=61906) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 854.816110] env[61906]: DEBUG nova.compute.manager [req-92302c58-ea9e-4483-bf49-bfcab0c1f201 req-b7d49345-0db5-41ff-9511-4d388661ce69 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Received event network-vif-plugged-a88ac5a5-d8da-4ebb-a745-ecc89275cb78 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.816362] env[61906]: DEBUG oslo_concurrency.lockutils [req-92302c58-ea9e-4483-bf49-bfcab0c1f201 req-b7d49345-0db5-41ff-9511-4d388661ce69 service nova] Acquiring lock "33e288d5-9065-4606-b17a-c4c90bcbd533-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.816581] env[61906]: DEBUG oslo_concurrency.lockutils [req-92302c58-ea9e-4483-bf49-bfcab0c1f201 req-b7d49345-0db5-41ff-9511-4d388661ce69 service nova] Lock "33e288d5-9065-4606-b17a-c4c90bcbd533-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.816757] env[61906]: DEBUG oslo_concurrency.lockutils [req-92302c58-ea9e-4483-bf49-bfcab0c1f201 req-b7d49345-0db5-41ff-9511-4d388661ce69 service nova] Lock "33e288d5-9065-4606-b17a-c4c90bcbd533-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.816945] env[61906]: DEBUG nova.compute.manager [req-92302c58-ea9e-4483-bf49-bfcab0c1f201 req-b7d49345-0db5-41ff-9511-4d388661ce69 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] No waiting events found dispatching network-vif-plugged-a88ac5a5-d8da-4ebb-a745-ecc89275cb78 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.817441] env[61906]: WARNING nova.compute.manager [req-92302c58-ea9e-4483-bf49-bfcab0c1f201 req-b7d49345-0db5-41ff-9511-4d388661ce69 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Received unexpected event network-vif-plugged-a88ac5a5-d8da-4ebb-a745-ecc89275cb78 for instance with vm_state building and task_state spawning. [ 854.913286] env[61906]: DEBUG oslo_vmware.api [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356721, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.283141} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.913496] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.913696] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 854.913971] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 854.914247] env[61906]: INFO nova.compute.manager [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Took 0.66 seconds to destroy the instance on the hypervisor. [ 854.914877] env[61906]: DEBUG oslo.service.loopingcall [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.914877] env[61906]: DEBUG nova.compute.manager [-] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 854.914877] env[61906]: DEBUG nova.network.neutron [-] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 854.939101] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.939650] env[61906]: DEBUG nova.compute.manager [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 855.021678] env[61906]: WARNING nova.network.neutron [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] b12c63e1-bd61-4da3-8d37-4a89206ba0f2 already exists in list: networks containing: ['b12c63e1-bd61-4da3-8d37-4a89206ba0f2']. ignoring it [ 855.055933] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "bb53f018-8546-4a00-a4be-d3c48416b1a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.056264] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "bb53f018-8546-4a00-a4be-d3c48416b1a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.104712] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356722, 'name': Rename_Task, 'duration_secs': 0.377311} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.104997] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.105281] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a281d901-f781-44b2-ab5b-c202e3057a5b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.112609] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 855.112609] env[61906]: value = "task-1356724" [ 855.112609] env[61906]: _type = "Task" [ 855.112609] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.128889] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356724, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.152638] env[61906]: DEBUG nova.network.neutron [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Successfully updated port: a88ac5a5-d8da-4ebb-a745-ecc89275cb78 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 855.166586] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356723, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.172378] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356719, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.684422} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.173291] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 3a6da3bd-8e28-4c23-8660-f32f5c862d02/3a6da3bd-8e28-4c23-8660-f32f5c862d02.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 855.175018] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 855.175018] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2ac1a6e8-2558-466a-9dcd-01a74c64da8d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.183510] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 855.183510] env[61906]: value = "task-1356725" [ 855.183510] env[61906]: _type = "Task" [ 855.183510] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.196625] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356725, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.196625] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.196625] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.344278] env[61906]: DEBUG nova.network.neutron [-] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.444724] env[61906]: DEBUG nova.compute.utils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 855.446282] env[61906]: DEBUG nova.compute.manager [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 855.446448] env[61906]: DEBUG nova.network.neutron [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 855.508021] env[61906]: DEBUG nova.policy [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '231e62b44cd94249ab22ef8c43e60c96', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24249dfdffc046ddb41d0da91203fa4a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 855.561406] env[61906]: DEBUG nova.compute.manager [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 855.626443] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356724, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.628847] env[61906]: DEBUG nova.compute.manager [req-345440ee-e7ac-4178-abdf-3149d5b0dc64 req-d2e9eba1-4708-4a70-bc87-cb69cee6f8d2 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Received event network-changed-a88ac5a5-d8da-4ebb-a745-ecc89275cb78 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.629183] env[61906]: DEBUG nova.compute.manager [req-345440ee-e7ac-4178-abdf-3149d5b0dc64 req-d2e9eba1-4708-4a70-bc87-cb69cee6f8d2 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Refreshing instance network info cache due to event network-changed-a88ac5a5-d8da-4ebb-a745-ecc89275cb78. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 855.629477] env[61906]: DEBUG oslo_concurrency.lockutils [req-345440ee-e7ac-4178-abdf-3149d5b0dc64 req-d2e9eba1-4708-4a70-bc87-cb69cee6f8d2 service nova] Acquiring lock "refresh_cache-33e288d5-9065-4606-b17a-c4c90bcbd533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.629692] env[61906]: DEBUG oslo_concurrency.lockutils [req-345440ee-e7ac-4178-abdf-3149d5b0dc64 req-d2e9eba1-4708-4a70-bc87-cb69cee6f8d2 service nova] Acquired lock "refresh_cache-33e288d5-9065-4606-b17a-c4c90bcbd533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.629863] env[61906]: DEBUG nova.network.neutron [req-345440ee-e7ac-4178-abdf-3149d5b0dc64 req-d2e9eba1-4708-4a70-bc87-cb69cee6f8d2 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Refreshing network info cache for port a88ac5a5-d8da-4ebb-a745-ecc89275cb78 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 855.658577] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "refresh_cache-33e288d5-9065-4606-b17a-c4c90bcbd533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.669361] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356723, 'name': ReconfigVM_Task, 'duration_secs': 0.628296} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.669675] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44/31d0f640-47f4-4b72-b9fe-6ef4c76c6b44.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.670364] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0ab607ac-76df-469f-9930-089eb0643d5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.678344] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 855.678344] env[61906]: value = "task-1356726" [ 855.678344] env[61906]: _type = "Task" [ 855.678344] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.698026] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356725, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091117} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.698673] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 855.701176] env[61906]: INFO nova.compute.claims [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.705262] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6faa02cd-eb86-4ae0-a667-484c588e7a73 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.728881] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 3a6da3bd-8e28-4c23-8660-f32f5c862d02/3a6da3bd-8e28-4c23-8660-f32f5c862d02.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.730282] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1f985fd-c3c2-427f-a830-2cf32bc93f31 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.752861] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 855.752861] env[61906]: value = "task-1356727" [ 855.752861] env[61906]: _type = "Task" [ 855.752861] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.763953] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356727, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.781647] env[61906]: DEBUG nova.network.neutron [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Updating instance_info_cache with network_info: [{"id": "76876157-42ef-4519-88b2-ca9a600e3b41", "address": "fa:16:3e:66:6f:62", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76876157-42", "ovs_interfaceid": "76876157-42ef-4519-88b2-ca9a600e3b41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "38287aaf-c4d5-440a-b756-6502a2c7dc02", "address": "fa:16:3e:d8:5d:12", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38287aaf-c4", "ovs_interfaceid": "38287aaf-c4d5-440a-b756-6502a2c7dc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.807907] env[61906]: DEBUG nova.network.neutron [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Successfully created port: eb78ba42-034c-44d7-9dd3-4157cd702c8c {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.846804] env[61906]: INFO nova.compute.manager [-] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Took 1.45 seconds to deallocate network for instance. [ 855.952404] env[61906]: DEBUG nova.compute.manager [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 856.020791] env[61906]: DEBUG nova.compute.manager [req-74701245-20ad-440b-89f4-7c1a1de71dea req-16b133ed-9bad-4a27-a09f-81b4746c9da0 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Received event network-changed-38287aaf-c4d5-440a-b756-6502a2c7dc02 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.021008] env[61906]: DEBUG nova.compute.manager [req-74701245-20ad-440b-89f4-7c1a1de71dea req-16b133ed-9bad-4a27-a09f-81b4746c9da0 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Refreshing instance network info cache due to event network-changed-38287aaf-c4d5-440a-b756-6502a2c7dc02. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 856.021224] env[61906]: DEBUG oslo_concurrency.lockutils [req-74701245-20ad-440b-89f4-7c1a1de71dea req-16b133ed-9bad-4a27-a09f-81b4746c9da0 service nova] Acquiring lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.080895] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.104992] env[61906]: DEBUG nova.network.neutron [-] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.126837] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356724, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.164288] env[61906]: DEBUG nova.network.neutron [req-345440ee-e7ac-4178-abdf-3149d5b0dc64 req-d2e9eba1-4708-4a70-bc87-cb69cee6f8d2 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 856.189766] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356726, 'name': Rename_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.210184] env[61906]: INFO nova.compute.resource_tracker [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating resource usage from migration 0fb2c22f-a902-4138-8a6b-071a4253a8d9 [ 856.272106] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.285361] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.286224] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.286445] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.289783] env[61906]: DEBUG oslo_concurrency.lockutils [req-74701245-20ad-440b-89f4-7c1a1de71dea req-16b133ed-9bad-4a27-a09f-81b4746c9da0 service nova] Acquired lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.292167] env[61906]: DEBUG nova.network.neutron [req-74701245-20ad-440b-89f4-7c1a1de71dea req-16b133ed-9bad-4a27-a09f-81b4746c9da0 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Refreshing network info cache for port 38287aaf-c4d5-440a-b756-6502a2c7dc02 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 856.296977] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830aebaa-c172-4d61-8e62-d0f3be7d3824 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.297744] env[61906]: DEBUG nova.network.neutron [req-345440ee-e7ac-4178-abdf-3149d5b0dc64 req-d2e9eba1-4708-4a70-bc87-cb69cee6f8d2 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.316297] env[61906]: DEBUG nova.virt.hardware [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.316587] env[61906]: DEBUG nova.virt.hardware [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.316752] env[61906]: DEBUG nova.virt.hardware [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.316936] env[61906]: DEBUG nova.virt.hardware [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.317110] env[61906]: DEBUG nova.virt.hardware [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.317270] env[61906]: DEBUG nova.virt.hardware [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.317471] env[61906]: DEBUG nova.virt.hardware [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.317631] env[61906]: DEBUG nova.virt.hardware [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.317801] env[61906]: DEBUG nova.virt.hardware [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.317970] env[61906]: DEBUG nova.virt.hardware [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.318205] env[61906]: DEBUG nova.virt.hardware [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.324386] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Reconfiguring VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 856.328115] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6df6157f-4c3b-4346-838b-412ab1739916 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.346678] env[61906]: DEBUG oslo_vmware.api [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 856.346678] env[61906]: value = "task-1356728" [ 856.346678] env[61906]: _type = "Task" [ 856.346678] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.352940] env[61906]: DEBUG oslo_concurrency.lockutils [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.359645] env[61906]: DEBUG oslo_vmware.api [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356728, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.510683] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e5dfd3-c2c4-4f2d-a56e-df8f58167102 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.520948] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ba2c3a-5aef-4719-b933-046bc77a24b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.526504] env[61906]: DEBUG oslo_concurrency.lockutils [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.526736] env[61906]: DEBUG oslo_concurrency.lockutils [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.554625] env[61906]: DEBUG nova.compute.utils [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 856.556678] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ceda57-0d40-45ec-9320-8987c932fa15 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.566297] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db71b93-01b5-48e6-b6d2-3b9893efc36d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.581764] env[61906]: DEBUG nova.compute.provider_tree [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.607632] env[61906]: INFO nova.compute.manager [-] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Took 1.69 seconds to deallocate network for instance. [ 856.626533] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356724, 'name': PowerOnVM_Task, 'duration_secs': 1.266898} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.626836] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.627013] env[61906]: INFO nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Took 10.14 seconds to spawn the instance on the hypervisor. [ 856.627245] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.628062] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9019d5-56f2-43ab-9453-7df5eef1bf85 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.690887] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356726, 'name': Rename_Task, 'duration_secs': 0.594161} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.691180] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.691435] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30dcf32a-f4da-4bdb-8d0b-5677423516b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.699015] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 856.699015] env[61906]: value = "task-1356729" [ 856.699015] env[61906]: _type = "Task" [ 856.699015] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.706914] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356729, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.763745] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356727, 'name': ReconfigVM_Task, 'duration_secs': 0.735066} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.764047] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 3a6da3bd-8e28-4c23-8660-f32f5c862d02/3a6da3bd-8e28-4c23-8660-f32f5c862d02.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.764706] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77b5334d-3e2d-4653-bde6-b31058db494f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.771633] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 856.771633] env[61906]: value = "task-1356730" [ 856.771633] env[61906]: _type = "Task" [ 856.771633] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.779454] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356730, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.803023] env[61906]: DEBUG oslo_concurrency.lockutils [req-345440ee-e7ac-4178-abdf-3149d5b0dc64 req-d2e9eba1-4708-4a70-bc87-cb69cee6f8d2 service nova] Releasing lock "refresh_cache-33e288d5-9065-4606-b17a-c4c90bcbd533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.803587] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "refresh_cache-33e288d5-9065-4606-b17a-c4c90bcbd533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.803936] env[61906]: DEBUG nova.network.neutron [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 856.857616] env[61906]: DEBUG oslo_vmware.api [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356728, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.964366] env[61906]: DEBUG nova.compute.manager [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 856.992904] env[61906]: DEBUG nova.virt.hardware [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.993181] env[61906]: DEBUG nova.virt.hardware [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.993351] env[61906]: DEBUG nova.virt.hardware [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.993542] env[61906]: DEBUG nova.virt.hardware [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.993738] env[61906]: DEBUG nova.virt.hardware [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.993909] env[61906]: DEBUG nova.virt.hardware [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.994236] env[61906]: DEBUG nova.virt.hardware [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.994426] env[61906]: DEBUG nova.virt.hardware [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.994598] env[61906]: DEBUG nova.virt.hardware [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.994766] env[61906]: DEBUG nova.virt.hardware [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.994947] env[61906]: DEBUG nova.virt.hardware [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.995808] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932d4cad-5fd7-4a31-aafb-4f7c1d03763e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.004031] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-675adccb-8097-4b58-97f2-76cef7ff238d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.021435] env[61906]: DEBUG nova.network.neutron [req-74701245-20ad-440b-89f4-7c1a1de71dea req-16b133ed-9bad-4a27-a09f-81b4746c9da0 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Updated VIF entry in instance network info cache for port 38287aaf-c4d5-440a-b756-6502a2c7dc02. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 857.021849] env[61906]: DEBUG nova.network.neutron [req-74701245-20ad-440b-89f4-7c1a1de71dea req-16b133ed-9bad-4a27-a09f-81b4746c9da0 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Updating instance_info_cache with network_info: [{"id": "76876157-42ef-4519-88b2-ca9a600e3b41", "address": "fa:16:3e:66:6f:62", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76876157-42", "ovs_interfaceid": "76876157-42ef-4519-88b2-ca9a600e3b41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "38287aaf-c4d5-440a-b756-6502a2c7dc02", "address": "fa:16:3e:d8:5d:12", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38287aaf-c4", "ovs_interfaceid": "38287aaf-c4d5-440a-b756-6502a2c7dc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.057074] env[61906]: DEBUG oslo_concurrency.lockutils [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 0.530s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.084586] env[61906]: DEBUG nova.scheduler.client.report [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.113759] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.143335] env[61906]: INFO nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Took 22.54 seconds to build instance. [ 857.212747] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356729, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.283284] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356730, 'name': Rename_Task, 'duration_secs': 0.191981} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.283591] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 857.283839] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-48b26477-fba2-4ccd-b0bf-f93aff312b7a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.292122] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 857.292122] env[61906]: value = "task-1356731" [ 857.292122] env[61906]: _type = "Task" [ 857.292122] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.300662] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356731, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.301520] env[61906]: DEBUG nova.network.neutron [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Successfully updated port: eb78ba42-034c-44d7-9dd3-4157cd702c8c {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.334471] env[61906]: DEBUG nova.network.neutron [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 857.359559] env[61906]: DEBUG oslo_vmware.api [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356728, 'name': ReconfigVM_Task, 'duration_secs': 0.833422} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.359911] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.360149] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Reconfigured VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 857.471505] env[61906]: DEBUG nova.network.neutron [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Updating instance_info_cache with network_info: [{"id": "a88ac5a5-d8da-4ebb-a745-ecc89275cb78", "address": "fa:16:3e:6e:19:ce", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa88ac5a5-d8", "ovs_interfaceid": "a88ac5a5-d8da-4ebb-a745-ecc89275cb78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.525195] env[61906]: DEBUG oslo_concurrency.lockutils [req-74701245-20ad-440b-89f4-7c1a1de71dea req-16b133ed-9bad-4a27-a09f-81b4746c9da0 service nova] Releasing lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.525195] env[61906]: DEBUG nova.compute.manager [req-74701245-20ad-440b-89f4-7c1a1de71dea req-16b133ed-9bad-4a27-a09f-81b4746c9da0 service nova] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Received event network-vif-deleted-fcbc3598-49e0-4da8-91e1-e0f6929f305f {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.525195] env[61906]: DEBUG nova.compute.manager [req-74701245-20ad-440b-89f4-7c1a1de71dea req-16b133ed-9bad-4a27-a09f-81b4746c9da0 service nova] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Received event network-vif-deleted-b670d2ba-324d-4d95-b044-de1cc7c68bd5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.525524] env[61906]: INFO nova.compute.manager [req-74701245-20ad-440b-89f4-7c1a1de71dea req-16b133ed-9bad-4a27-a09f-81b4746c9da0 service nova] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Neutron deleted interface b670d2ba-324d-4d95-b044-de1cc7c68bd5; detaching it from the instance and deleting it from the info cache [ 857.525524] env[61906]: DEBUG nova.network.neutron [req-74701245-20ad-440b-89f4-7c1a1de71dea req-16b133ed-9bad-4a27-a09f-81b4746c9da0 service nova] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.590681] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.394s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.590873] env[61906]: INFO nova.compute.manager [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Migrating [ 857.591115] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.591269] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.592754] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.512s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.594403] env[61906]: INFO nova.compute.claims [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.645863] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "63cbf664-078b-40c5-92bb-7a5a8d8aea3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.054s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.711571] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356729, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.803769] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.804145] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.804515] env[61906]: DEBUG nova.network.neutron [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 857.805589] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356731, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.865770] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7273ba5-2545-4284-bc84-289659718bb2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-5adaa660-b736-4c11-9141-846cf475ccd5-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.651s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.974834] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "refresh_cache-33e288d5-9065-4606-b17a-c4c90bcbd533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.975183] env[61906]: DEBUG nova.compute.manager [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Instance network_info: |[{"id": "a88ac5a5-d8da-4ebb-a745-ecc89275cb78", "address": "fa:16:3e:6e:19:ce", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa88ac5a5-d8", "ovs_interfaceid": "a88ac5a5-d8da-4ebb-a745-ecc89275cb78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 857.975621] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:19:ce', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30f1dacf-8988-41b8-aa8f-e9530f65ef46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a88ac5a5-d8da-4ebb-a745-ecc89275cb78', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 857.983217] env[61906]: DEBUG oslo.service.loopingcall [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.983528] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 857.983697] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-07ecdaf9-40b0-4936-8432-8557d79deb67 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.004507] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.004507] env[61906]: value = "task-1356732" [ 858.004507] env[61906]: _type = "Task" [ 858.004507] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.014744] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356732, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.028638] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fc7b1d87-deee-41bc-8939-ecd53cb3477e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.042239] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f6ff749-f944-4479-b5e4-92ada9a44ef8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.074907] env[61906]: DEBUG nova.compute.manager [req-74701245-20ad-440b-89f4-7c1a1de71dea req-16b133ed-9bad-4a27-a09f-81b4746c9da0 service nova] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Detach interface failed, port_id=b670d2ba-324d-4d95-b044-de1cc7c68bd5, reason: Instance 24b4a747-67aa-4388-aed6-cb646cd55765 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 858.093889] env[61906]: INFO nova.compute.rpcapi [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 858.095076] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.101279] env[61906]: DEBUG oslo_concurrency.lockutils [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.101520] env[61906]: DEBUG oslo_concurrency.lockutils [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.101810] env[61906]: INFO nova.compute.manager [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Attaching volume 085526e0-9c7f-46ce-9f8b-d76aa54f673e to /dev/sdb [ 858.142875] env[61906]: DEBUG nova.compute.manager [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Received event network-vif-plugged-eb78ba42-034c-44d7-9dd3-4157cd702c8c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.143206] env[61906]: DEBUG oslo_concurrency.lockutils [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] Acquiring lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.143740] env[61906]: DEBUG oslo_concurrency.lockutils [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.143740] env[61906]: DEBUG oslo_concurrency.lockutils [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.143740] env[61906]: DEBUG nova.compute.manager [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] No waiting events found dispatching network-vif-plugged-eb78ba42-034c-44d7-9dd3-4157cd702c8c {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 858.143926] env[61906]: WARNING nova.compute.manager [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Received unexpected event network-vif-plugged-eb78ba42-034c-44d7-9dd3-4157cd702c8c for instance with vm_state building and task_state spawning. [ 858.144143] env[61906]: DEBUG nova.compute.manager [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Received event network-changed-eb78ba42-034c-44d7-9dd3-4157cd702c8c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.144268] env[61906]: DEBUG nova.compute.manager [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Refreshing instance network info cache due to event network-changed-eb78ba42-034c-44d7-9dd3-4157cd702c8c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.144440] env[61906]: DEBUG oslo_concurrency.lockutils [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] Acquiring lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.146564] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65531e2a-ee72-48cd-bda0-9391fd50dfea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.154938] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358780cb-c4d7-402b-b656-b33fe30ce969 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.170455] env[61906]: DEBUG nova.virt.block_device [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Updating existing volume attachment record: bdec98be-bc7e-4b19-9d05-adf1e00f70ee {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 858.212101] env[61906]: DEBUG oslo_vmware.api [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356729, 'name': PowerOnVM_Task, 'duration_secs': 1.038019} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.213757] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 858.214055] env[61906]: DEBUG nova.compute.manager [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.215204] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60881192-f059-413b-aa75-88a70a6872b1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.302210] env[61906]: DEBUG oslo_vmware.api [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356731, 'name': PowerOnVM_Task, 'duration_secs': 0.979701} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.302499] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 858.302713] env[61906]: INFO nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Took 9.45 seconds to spawn the instance on the hypervisor. [ 858.302972] env[61906]: DEBUG nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.303769] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad54b95b-e767-42b7-8828-3373026b6d2d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.345312] env[61906]: DEBUG nova.network.neutron [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.501225] env[61906]: DEBUG nova.network.neutron [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updating instance_info_cache with network_info: [{"id": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "address": "fa:16:3e:f0:1b:68", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb78ba42-03", "ovs_interfaceid": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.515084] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356732, 'name': CreateVM_Task, 'duration_secs': 0.41424} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.515872] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 858.516717] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.516927] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.517551] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 858.518100] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59ce3d89-628a-4013-8894-69914d8ff3cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.523452] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 858.523452] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52268c90-e8cc-7aab-8970-eb559468b2c2" [ 858.523452] env[61906]: _type = "Task" [ 858.523452] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.532323] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52268c90-e8cc-7aab-8970-eb559468b2c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.613905] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.613905] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.614227] env[61906]: DEBUG nova.network.neutron [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 858.735198] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.826698] env[61906]: INFO nova.compute.manager [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Took 24.19 seconds to build instance. [ 858.872260] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-648444ac-14ae-4385-b7e4-b874ec8fec97 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.882191] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf64be17-9057-4773-ad65-bddbba1dba91 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.912797] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664599a3-f653-4a50-8f9d-d82c2e2f8032 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.922087] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6767cc9-f444-42c6-ab71-a3098d3c30c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.937600] env[61906]: DEBUG nova.compute.provider_tree [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.003968] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.004384] env[61906]: DEBUG nova.compute.manager [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Instance network_info: |[{"id": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "address": "fa:16:3e:f0:1b:68", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb78ba42-03", "ovs_interfaceid": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 859.004731] env[61906]: DEBUG oslo_concurrency.lockutils [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] Acquired lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.004955] env[61906]: DEBUG nova.network.neutron [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Refreshing network info cache for port eb78ba42-034c-44d7-9dd3-4157cd702c8c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 859.006506] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:1b:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '680cb499-2a47-482b-af0d-112016ac0e17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb78ba42-034c-44d7-9dd3-4157cd702c8c', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.014326] env[61906]: DEBUG oslo.service.loopingcall [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.017661] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 859.020684] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bc2ba8b3-f088-4281-a64f-a562d3818fcb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.046204] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52268c90-e8cc-7aab-8970-eb559468b2c2, 'name': SearchDatastore_Task, 'duration_secs': 0.010468} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.047571] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.047827] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.048108] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.048283] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.048476] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.048824] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.048824] env[61906]: value = "task-1356736" [ 859.048824] env[61906]: _type = "Task" [ 859.048824] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.049213] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b5e29966-0663-4164-82ae-e385a6e3a0cc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.060432] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356736, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.061573] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.061815] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 859.062482] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f372e43c-adea-4baf-a5e0-541de69f5414 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.070985] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 859.070985] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a99b1b-8ddb-dc67-8080-4929d3aa5145" [ 859.070985] env[61906]: _type = "Task" [ 859.070985] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.081410] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a99b1b-8ddb-dc67-8080-4929d3aa5145, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.328848] env[61906]: DEBUG oslo_concurrency.lockutils [None req-8f5408a4-f419-4aad-a325-38b0cdeb7d0d tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "3a6da3bd-8e28-4c23-8660-f32f5c862d02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.705s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.345883] env[61906]: DEBUG nova.network.neutron [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updated VIF entry in instance network info cache for port eb78ba42-034c-44d7-9dd3-4157cd702c8c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.345883] env[61906]: DEBUG nova.network.neutron [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updating instance_info_cache with network_info: [{"id": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "address": "fa:16:3e:f0:1b:68", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb78ba42-03", "ovs_interfaceid": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.437876] env[61906]: DEBUG nova.network.neutron [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance_info_cache with network_info: [{"id": "ba127004-dc19-4fc4-b84b-97584ac68f34", "address": "fa:16:3e:d1:ad:74", "network": {"id": "9e0705ba-7b9f-46fb-8603-01598a888212", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-705593104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb85e88fd4f54e0db2ff131f81137f64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba127004-dc", "ovs_interfaceid": "ba127004-dc19-4fc4-b84b-97584ac68f34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.440529] env[61906]: DEBUG nova.scheduler.client.report [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.561338] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356736, 'name': CreateVM_Task, 'duration_secs': 0.347} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.561547] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.562238] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.562409] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.562726] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.562987] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d5516ff-a0df-4399-8ef1-55fea89c4a2f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.571411] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 859.571411] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52886d4e-3ca5-9956-c5fe-27bacd7beea8" [ 859.571411] env[61906]: _type = "Task" [ 859.571411] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.585088] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a99b1b-8ddb-dc67-8080-4929d3aa5145, 'name': SearchDatastore_Task, 'duration_secs': 0.01241} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.588858] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52886d4e-3ca5-9956-c5fe-27bacd7beea8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.589116] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d1b7e4f-5fe7-4f7f-8ab7-27e9d273be18 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.595625] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 859.595625] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a8402d-07a0-98b5-43f7-b2968220241a" [ 859.595625] env[61906]: _type = "Task" [ 859.595625] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.606295] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a8402d-07a0-98b5-43f7-b2968220241a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.744176] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "interface-5adaa660-b736-4c11-9141-846cf475ccd5-38287aaf-c4d5-440a-b756-6502a2c7dc02" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.744526] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-5adaa660-b736-4c11-9141-846cf475ccd5-38287aaf-c4d5-440a-b756-6502a2c7dc02" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.847968] env[61906]: DEBUG oslo_concurrency.lockutils [req-eb0cac13-2a0d-4508-96ce-c45cf1f0c1cb req-2e268ebf-2fd4-4950-aec7-c987fa986da0 service nova] Releasing lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.880023] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "31d0f640-47f4-4b72-b9fe-6ef4c76c6b44" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.880342] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "31d0f640-47f4-4b72-b9fe-6ef4c76c6b44" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.880573] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "31d0f640-47f4-4b72-b9fe-6ef4c76c6b44-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.880871] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "31d0f640-47f4-4b72-b9fe-6ef4c76c6b44-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.881202] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "31d0f640-47f4-4b72-b9fe-6ef4c76c6b44-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.883550] env[61906]: INFO nova.compute.manager [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Terminating instance [ 859.886039] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "refresh_cache-31d0f640-47f4-4b72-b9fe-6ef4c76c6b44" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.886284] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquired lock "refresh_cache-31d0f640-47f4-4b72-b9fe-6ef4c76c6b44" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.886499] env[61906]: DEBUG nova.network.neutron [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 859.947919] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.949863] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.950454] env[61906]: DEBUG nova.compute.manager [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 859.952970] env[61906]: DEBUG oslo_concurrency.lockutils [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.600s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.953207] env[61906]: DEBUG nova.objects.instance [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Lazy-loading 'resources' on Instance uuid 0af2b07b-0632-4376-b502-58fe23d79f02 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.069753] env[61906]: DEBUG oslo_concurrency.lockutils [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "63cbf664-078b-40c5-92bb-7a5a8d8aea3b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.070185] env[61906]: DEBUG oslo_concurrency.lockutils [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "63cbf664-078b-40c5-92bb-7a5a8d8aea3b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.070546] env[61906]: DEBUG oslo_concurrency.lockutils [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "63cbf664-078b-40c5-92bb-7a5a8d8aea3b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.070844] env[61906]: DEBUG oslo_concurrency.lockutils [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "63cbf664-078b-40c5-92bb-7a5a8d8aea3b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.071123] env[61906]: DEBUG oslo_concurrency.lockutils [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "63cbf664-078b-40c5-92bb-7a5a8d8aea3b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.074101] env[61906]: INFO nova.compute.manager [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Terminating instance [ 860.080734] env[61906]: DEBUG nova.compute.manager [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 860.081036] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 860.082152] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3cbc5e5-d0e2-4239-9700-61ce21d7331d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.094527] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52886d4e-3ca5-9956-c5fe-27bacd7beea8, 'name': SearchDatastore_Task, 'duration_secs': 0.024556} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.097515] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.097844] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.098172] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.098636] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 860.102798] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4010292-187e-4a9b-a923-656d0d666dec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.112490] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a8402d-07a0-98b5-43f7-b2968220241a, 'name': SearchDatastore_Task, 'duration_secs': 0.025749} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.114013] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.114396] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 33e288d5-9065-4606-b17a-c4c90bcbd533/33e288d5-9065-4606-b17a-c4c90bcbd533.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 860.114842] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 860.114842] env[61906]: value = "task-1356737" [ 860.114842] env[61906]: _type = "Task" [ 860.114842] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.115160] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.115451] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.115745] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d99f13c-d1c9-4cb5-a7ad-a47758751024 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.118421] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a77ccd01-e545-44c1-adfe-b26cf5660140 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.132699] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.134680] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.134873] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.135739] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "3a6da3bd-8e28-4c23-8660-f32f5c862d02" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.135974] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "3a6da3bd-8e28-4c23-8660-f32f5c862d02" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.136256] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "3a6da3bd-8e28-4c23-8660-f32f5c862d02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.136471] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "3a6da3bd-8e28-4c23-8660-f32f5c862d02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.136661] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "3a6da3bd-8e28-4c23-8660-f32f5c862d02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.138395] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 860.138395] env[61906]: value = "task-1356738" [ 860.138395] env[61906]: _type = "Task" [ 860.138395] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.138822] env[61906]: INFO nova.compute.manager [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Terminating instance [ 860.140086] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71758603-e4cb-437c-841c-6d710c3b4414 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.142789] env[61906]: DEBUG nova.compute.manager [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 860.142977] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 860.146442] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c731970-c611-42c4-a828-7ff7cf956ce2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.151962] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 860.151962] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b7cef7-05fa-1749-5498-8f319a11894b" [ 860.151962] env[61906]: _type = "Task" [ 860.151962] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.154924] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356738, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.159656] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 860.159879] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-983d9f67-7271-411b-9857-3a75d62213da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.166721] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b7cef7-05fa-1749-5498-8f319a11894b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.168010] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 860.168010] env[61906]: value = "task-1356739" [ 860.168010] env[61906]: _type = "Task" [ 860.168010] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.176754] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356739, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.247798] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.248070] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.249101] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab4d98d-8b9b-475f-9345-b3e3b844fbe6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.269817] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e52f94-d9a7-4673-bbc0-1bffcf449883 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.296351] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Reconfiguring VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 860.296606] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3060a26-c143-4d70-963b-526ae6d01416 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.317085] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 860.317085] env[61906]: value = "task-1356740" [ 860.317085] env[61906]: _type = "Task" [ 860.317085] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.329416] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.409363] env[61906]: DEBUG nova.network.neutron [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.460554] env[61906]: DEBUG nova.compute.utils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.465651] env[61906]: DEBUG nova.compute.manager [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 860.465905] env[61906]: DEBUG nova.network.neutron [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 860.511057] env[61906]: DEBUG nova.network.neutron [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.578408] env[61906]: DEBUG nova.policy [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ec22db431334e7886acb497ebfc7eb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ac9a6686e7184fb1913e84a77985b449', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 860.635531] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.653767] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356738, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.670565] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b7cef7-05fa-1749-5498-8f319a11894b, 'name': SearchDatastore_Task, 'duration_secs': 0.017665} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.674752] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f32c6ed2-9858-409e-b33d-e53af93ea13a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.684055] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 860.684055] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5229d4af-2a68-08a4-d957-66cb86dd862c" [ 860.684055] env[61906]: _type = "Task" [ 860.684055] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.687456] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356739, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.698502] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5229d4af-2a68-08a4-d957-66cb86dd862c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.791458] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e4e579a-402c-428a-8aa3-899c9adf89de {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.801773] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba193a3-54b9-4b91-ac2f-0a6546e6dccf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.835930] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef8ee305-9889-46e3-9ec4-f5774f34184a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.845370] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.849160] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1279fae-8ae4-4042-b075-09e003ccae6c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.864603] env[61906]: DEBUG nova.compute.provider_tree [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.917045] env[61906]: DEBUG nova.network.neutron [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Successfully created port: 7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 860.970934] env[61906]: DEBUG nova.compute.manager [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 861.014140] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Releasing lock "refresh_cache-31d0f640-47f4-4b72-b9fe-6ef4c76c6b44" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.015277] env[61906]: DEBUG nova.compute.manager [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 861.015961] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 861.018926] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea90af7c-0742-46ba-bbce-b72b90ef739a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.030980] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 861.031707] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c9482dc0-5dc6-46b9-a42b-ba4a95361f52 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.041754] env[61906]: DEBUG oslo_vmware.api [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 861.041754] env[61906]: value = "task-1356742" [ 861.041754] env[61906]: _type = "Task" [ 861.041754] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.052530] env[61906]: DEBUG oslo_vmware.api [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356742, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.131954] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.155505] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356738, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575077} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.155889] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 33e288d5-9065-4606-b17a-c4c90bcbd533/33e288d5-9065-4606-b17a-c4c90bcbd533.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 861.156218] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.156531] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2e25f64-87d7-4bf8-ae99-72cec2a16963 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.165766] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 861.165766] env[61906]: value = "task-1356743" [ 861.165766] env[61906]: _type = "Task" [ 861.165766] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.176827] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356743, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.182684] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356739, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.198128] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5229d4af-2a68-08a4-d957-66cb86dd862c, 'name': SearchDatastore_Task, 'duration_secs': 0.025552} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.198419] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.198719] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63/7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.199023] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50f65d3d-0993-407c-abc8-63403155aa7e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.206302] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 861.206302] env[61906]: value = "task-1356744" [ 861.206302] env[61906]: _type = "Task" [ 861.206302] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.215730] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356744, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.343396] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.373500] env[61906]: DEBUG nova.scheduler.client.report [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.480988] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3d650c-e93f-408f-aec5-bee20a96cc3e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.513316] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance '943c57f9-74c6-4465-9773-3fb01e78127b' progress to 0 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 861.554703] env[61906]: DEBUG oslo_vmware.api [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356742, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.635512] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.685211] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356739, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.688952] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356743, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110063} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.689300] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.690345] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6e14ac-4a07-481c-809f-903451e96e58 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.718735] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 33e288d5-9065-4606-b17a-c4c90bcbd533/33e288d5-9065-4606-b17a-c4c90bcbd533.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.722373] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d434cbb-4ed1-4015-ad18-29a03df89f2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.746439] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356744, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.748140] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 861.748140] env[61906]: value = "task-1356745" [ 861.748140] env[61906]: _type = "Task" [ 861.748140] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.757801] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356745, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.843251] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.877420] env[61906]: DEBUG oslo_concurrency.lockutils [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.924s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.880092] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.766s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.880359] env[61906]: DEBUG nova.objects.instance [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lazy-loading 'resources' on Instance uuid 24b4a747-67aa-4388-aed6-cb646cd55765 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.901601] env[61906]: INFO nova.scheduler.client.report [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Deleted allocations for instance 0af2b07b-0632-4376-b502-58fe23d79f02 [ 861.987694] env[61906]: DEBUG nova.compute.manager [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 862.015390] env[61906]: DEBUG nova.virt.hardware [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.015695] env[61906]: DEBUG nova.virt.hardware [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.015862] env[61906]: DEBUG nova.virt.hardware [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.016068] env[61906]: DEBUG nova.virt.hardware [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.016299] env[61906]: DEBUG nova.virt.hardware [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.016438] env[61906]: DEBUG nova.virt.hardware [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.016655] env[61906]: DEBUG nova.virt.hardware [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.016827] env[61906]: DEBUG nova.virt.hardware [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.017047] env[61906]: DEBUG nova.virt.hardware [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.017233] env[61906]: DEBUG nova.virt.hardware [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.017414] env[61906]: DEBUG nova.virt.hardware [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.018313] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b868ce-46d2-4e76-8736-7f0bf1c2f9ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.022865] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 862.023476] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a176d3f-7057-433b-8420-73f14f4608ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.033046] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809681fc-45b5-48ba-809c-d103dd748059 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.037916] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 862.037916] env[61906]: value = "task-1356746" [ 862.037916] env[61906]: _type = "Task" [ 862.037916] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.056567] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356746, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.061704] env[61906]: DEBUG oslo_vmware.api [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356742, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.135103] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.182015] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356739, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.220328] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356744, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.591818} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.220605] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63/7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 862.220824] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.221088] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e7823840-0449-4b7d-9f63-9187ed364842 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.228871] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 862.228871] env[61906]: value = "task-1356747" [ 862.228871] env[61906]: _type = "Task" [ 862.228871] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.238741] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356747, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.263033] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356745, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.343828] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.409574] env[61906]: DEBUG oslo_concurrency.lockutils [None req-041093d7-e126-4643-9d20-24f0de721177 tempest-ServersNegativeTestMultiTenantJSON-465680862 tempest-ServersNegativeTestMultiTenantJSON-465680862-project-member] Lock "0af2b07b-0632-4376-b502-58fe23d79f02" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.233s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.551103] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356746, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.562906] env[61906]: DEBUG oslo_vmware.api [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356742, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.626165] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bcd0054-a5b8-4e8a-9db6-631d755c6245 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.640809] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36b79c4d-8911-4762-9784-63c386c86e09 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.643903] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.673017] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7cdb11-8982-4966-a0db-5d425cc699cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.676524] env[61906]: DEBUG nova.compute.manager [req-035a820e-60cb-4237-9658-e086526e8139 req-51a43341-a75e-4edc-a3a4-82c142a65ab5 service nova] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Received event network-vif-plugged-7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.676737] env[61906]: DEBUG oslo_concurrency.lockutils [req-035a820e-60cb-4237-9658-e086526e8139 req-51a43341-a75e-4edc-a3a4-82c142a65ab5 service nova] Acquiring lock "bb53f018-8546-4a00-a4be-d3c48416b1a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.676947] env[61906]: DEBUG oslo_concurrency.lockutils [req-035a820e-60cb-4237-9658-e086526e8139 req-51a43341-a75e-4edc-a3a4-82c142a65ab5 service nova] Lock "bb53f018-8546-4a00-a4be-d3c48416b1a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.677133] env[61906]: DEBUG oslo_concurrency.lockutils [req-035a820e-60cb-4237-9658-e086526e8139 req-51a43341-a75e-4edc-a3a4-82c142a65ab5 service nova] Lock "bb53f018-8546-4a00-a4be-d3c48416b1a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.677303] env[61906]: DEBUG nova.compute.manager [req-035a820e-60cb-4237-9658-e086526e8139 req-51a43341-a75e-4edc-a3a4-82c142a65ab5 service nova] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] No waiting events found dispatching network-vif-plugged-7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 862.677474] env[61906]: WARNING nova.compute.manager [req-035a820e-60cb-4237-9658-e086526e8139 req-51a43341-a75e-4edc-a3a4-82c142a65ab5 service nova] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Received unexpected event network-vif-plugged-7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6 for instance with vm_state building and task_state spawning. [ 862.686941] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356739, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.689973] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407fb928-44d0-4fcf-97d7-19e23792c5d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.704681] env[61906]: DEBUG nova.compute.provider_tree [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.740120] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356747, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078409} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.741260] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 862.741434] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288996', 'volume_id': '085526e0-9c7f-46ce-9f8b-d76aa54f673e', 'name': 'volume-085526e0-9c7f-46ce-9f8b-d76aa54f673e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '51fa46a9-b1b3-4115-b1d7-cc4f672f2e83', 'attached_at': '', 'detached_at': '', 'volume_id': '085526e0-9c7f-46ce-9f8b-d76aa54f673e', 'serial': '085526e0-9c7f-46ce-9f8b-d76aa54f673e'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 862.741812] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.746018] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694d6c16-8740-4333-b7f6-306c584a9574 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.746018] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd55421c-9292-490c-9275-2d0b1648c3c7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.787703] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63/7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.789057] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6d32df0-bf8e-4a9f-8a82-0b63fd1c2fce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.804855] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa44546f-c493-4362-a102-f357286fb302 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.810831] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356745, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.835315] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] volume-085526e0-9c7f-46ce-9f8b-d76aa54f673e/volume-085526e0-9c7f-46ce-9f8b-d76aa54f673e.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.836968] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42ec61ff-bcec-4de7-ab93-375f7a3be79f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.849713] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 862.849713] env[61906]: value = "task-1356748" [ 862.849713] env[61906]: _type = "Task" [ 862.849713] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.859521] env[61906]: DEBUG oslo_vmware.api [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 862.859521] env[61906]: value = "task-1356749" [ 862.859521] env[61906]: _type = "Task" [ 862.859521] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.865242] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356748, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.865848] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.873981] env[61906]: DEBUG oslo_vmware.api [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356749, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.888250] env[61906]: DEBUG nova.network.neutron [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Successfully updated port: 7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 863.048718] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356746, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.059404] env[61906]: DEBUG oslo_vmware.api [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356742, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.135819] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.190107] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356739, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.212024] env[61906]: DEBUG nova.scheduler.client.report [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.261760] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356745, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.348478] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.361060] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356748, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.375413] env[61906]: DEBUG oslo_vmware.api [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356749, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.391112] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "refresh_cache-bb53f018-8546-4a00-a4be-d3c48416b1a1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.392494] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "refresh_cache-bb53f018-8546-4a00-a4be-d3c48416b1a1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.393483] env[61906]: DEBUG nova.network.neutron [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 863.552425] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356746, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.562552] env[61906]: DEBUG oslo_vmware.api [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356742, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.641389] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.690385] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356739, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.717498] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.837s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.719902] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 4.985s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.720123] env[61906]: DEBUG nova.objects.instance [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61906) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 863.760790] env[61906]: INFO nova.scheduler.client.report [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleted allocations for instance 24b4a747-67aa-4388-aed6-cb646cd55765 [ 863.766825] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356745, 'name': ReconfigVM_Task, 'duration_secs': 1.866894} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.767747] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 33e288d5-9065-4606-b17a-c4c90bcbd533/33e288d5-9065-4606-b17a-c4c90bcbd533.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.769323] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-598126f0-1e08-4223-9936-1ccf1db25faa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.777453] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 863.777453] env[61906]: value = "task-1356750" [ 863.777453] env[61906]: _type = "Task" [ 863.777453] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.788998] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356750, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.849648] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.868458] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356748, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.881379] env[61906]: DEBUG oslo_vmware.api [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356749, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.932451] env[61906]: DEBUG nova.network.neutron [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 864.051841] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356746, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.060742] env[61906]: DEBUG oslo_vmware.api [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356742, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.079829] env[61906]: DEBUG nova.network.neutron [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Updating instance_info_cache with network_info: [{"id": "7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6", "address": "fa:16:3e:01:e5:5b", "network": {"id": "5b29c9e0-23c9-42ee-9be4-60178bddedbe", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643906254-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac9a6686e7184fb1913e84a77985b449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c4d0cd2-9f", "ovs_interfaceid": "7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.137615] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356737, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.189590] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356739, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.271334] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41e55ab5-06d2-439b-bb4a-6dd26b70b960 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "24b4a747-67aa-4388-aed6-cb646cd55765" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.027s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.290484] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356750, 'name': Rename_Task, 'duration_secs': 0.180156} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.290484] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 864.290484] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-abaeb090-257f-4467-8dd7-403d4785901b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.297702] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 864.297702] env[61906]: value = "task-1356751" [ 864.297702] env[61906]: _type = "Task" [ 864.297702] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.069699] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "refresh_cache-bb53f018-8546-4a00-a4be-d3c48416b1a1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.070052] env[61906]: DEBUG nova.compute.manager [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Instance network_info: |[{"id": "7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6", "address": "fa:16:3e:01:e5:5b", "network": {"id": "5b29c9e0-23c9-42ee-9be4-60178bddedbe", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643906254-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac9a6686e7184fb1913e84a77985b449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c4d0cd2-9f", "ovs_interfaceid": "7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 865.071028] env[61906]: DEBUG oslo_concurrency.lockutils [None req-5a6e19ff-080a-4b51-b130-d6bd98ebb0e0 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.351s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.075064] env[61906]: DEBUG nova.compute.manager [req-8be7a708-c373-45fe-a075-7fd0717ce0ff req-38c4d74d-0c7a-45a5-9eb4-5e1d7922028f service nova] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Received event network-changed-7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.075064] env[61906]: DEBUG nova.compute.manager [req-8be7a708-c373-45fe-a075-7fd0717ce0ff req-38c4d74d-0c7a-45a5-9eb4-5e1d7922028f service nova] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Refreshing instance network info cache due to event network-changed-7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 865.075064] env[61906]: DEBUG oslo_concurrency.lockutils [req-8be7a708-c373-45fe-a075-7fd0717ce0ff req-38c4d74d-0c7a-45a5-9eb4-5e1d7922028f service nova] Acquiring lock "refresh_cache-bb53f018-8546-4a00-a4be-d3c48416b1a1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.075064] env[61906]: DEBUG oslo_concurrency.lockutils [req-8be7a708-c373-45fe-a075-7fd0717ce0ff req-38c4d74d-0c7a-45a5-9eb4-5e1d7922028f service nova] Acquired lock "refresh_cache-bb53f018-8546-4a00-a4be-d3c48416b1a1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.075064] env[61906]: DEBUG nova.network.neutron [req-8be7a708-c373-45fe-a075-7fd0717ce0ff req-38c4d74d-0c7a-45a5-9eb4-5e1d7922028f service nova] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Refreshing network info cache for port 7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 865.090323] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:e5:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c02dd284-ab80-451c-93eb-48c8360acb9c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.097719] env[61906]: DEBUG oslo.service.loopingcall [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.111311] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 865.119537] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-efce1439-b8d1-40ec-a9c6-ed7b1176b89a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.133874] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356751, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.134107] env[61906]: WARNING oslo_vmware.common.loopingcall [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] task run outlasted interval by 0.33647400000000005 sec [ 865.146295] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356748, 'name': ReconfigVM_Task, 'duration_secs': 1.108929} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.146987] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.157827] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63/7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.158676] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356737, 'name': PowerOffVM_Task, 'duration_secs': 4.579388} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.159043] env[61906]: DEBUG oslo_vmware.api [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356749, 'name': ReconfigVM_Task, 'duration_secs': 1.201906} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.168809] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6fab758-b99a-466b-b6d9-69cf4ca24215 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.170656] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 865.170756] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 865.171053] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Reconfigured VM instance instance-00000043 to attach disk [datastore2] volume-085526e0-9c7f-46ce-9f8b-d76aa54f673e/volume-085526e0-9c7f-46ce-9f8b-d76aa54f673e.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.175987] env[61906]: DEBUG oslo_vmware.api [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356742, 'name': PowerOffVM_Task, 'duration_secs': 3.642287} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.176373] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 865.176373] env[61906]: value = "task-1356752" [ 865.176373] env[61906]: _type = "Task" [ 865.176373] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.176619] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356739, 'name': PowerOffVM_Task, 'duration_secs': 4.523927} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.176841] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356746, 'name': PowerOffVM_Task, 'duration_secs': 2.658667} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.179628] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8531d392-8969-44cd-8a12-7c110f808f70 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.181537] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-643fe6c8-51ec-44bd-9887-dec9acb8b910 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.192144] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 865.192341] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 865.192671] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 865.192834] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 865.193077] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 865.193375] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance '943c57f9-74c6-4465-9773-3fb01e78127b' progress to 17 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 865.202219] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-50c6ec74-9015-4ee6-8484-60ca6f772f7e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.205987] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e84243a-4d35-42ae-94ea-51fd4b906e97 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.208080] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356751, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.210786] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 865.210786] env[61906]: value = "task-1356753" [ 865.210786] env[61906]: _type = "Task" [ 865.210786] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.219209] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356752, 'name': CreateVM_Task} progress is 15%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.221015] env[61906]: DEBUG oslo_vmware.api [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 865.221015] env[61906]: value = "task-1356755" [ 865.221015] env[61906]: _type = "Task" [ 865.221015] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.227688] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356753, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.235396] env[61906]: DEBUG oslo_vmware.api [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356755, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.241915] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 865.241915] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 865.241915] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Deleting the datastore file [datastore1] 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 865.241915] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-941a3f27-4a11-4594-a89e-a098fcb75881 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.253654] env[61906]: DEBUG oslo_vmware.api [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 865.253654] env[61906]: value = "task-1356758" [ 865.253654] env[61906]: _type = "Task" [ 865.253654] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.265327] env[61906]: DEBUG oslo_vmware.api [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356758, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.438962] env[61906]: DEBUG nova.network.neutron [req-8be7a708-c373-45fe-a075-7fd0717ce0ff req-38c4d74d-0c7a-45a5-9eb4-5e1d7922028f service nova] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Updated VIF entry in instance network info cache for port 7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 865.438962] env[61906]: DEBUG nova.network.neutron [req-8be7a708-c373-45fe-a075-7fd0717ce0ff req-38c4d74d-0c7a-45a5-9eb4-5e1d7922028f service nova] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Updating instance_info_cache with network_info: [{"id": "7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6", "address": "fa:16:3e:01:e5:5b", "network": {"id": "5b29c9e0-23c9-42ee-9be4-60178bddedbe", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643906254-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac9a6686e7184fb1913e84a77985b449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c4d0cd2-9f", "ovs_interfaceid": "7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.579515] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.657179] env[61906]: DEBUG oslo_vmware.api [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356751, 'name': PowerOnVM_Task, 'duration_secs': 0.903652} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.657416] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 865.657635] env[61906]: INFO nova.compute.manager [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Took 11.56 seconds to spawn the instance on the hypervisor. [ 865.657815] env[61906]: DEBUG nova.compute.manager [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.658639] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f71d23d-a98d-450c-ad8f-fca6daf55d71 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.713467] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 865.713773] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 865.713966] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 865.714299] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 865.714477] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 865.714632] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 865.714848] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 865.715029] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 865.715208] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 865.715391] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 865.715574] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 865.723072] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356752, 'name': CreateVM_Task, 'duration_secs': 0.535933} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.723598] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13351d23-b866-4e6b-8a6d-552e068e8290 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.735457] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 865.742249] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.742491] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.742868] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 865.743605] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-261a1fb5-480e-424b-af84-94e48b670d4c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.752204] env[61906]: DEBUG oslo_vmware.api [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356755, 'name': ReconfigVM_Task, 'duration_secs': 0.210948} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.756479] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288996', 'volume_id': '085526e0-9c7f-46ce-9f8b-d76aa54f673e', 'name': 'volume-085526e0-9c7f-46ce-9f8b-d76aa54f673e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '51fa46a9-b1b3-4115-b1d7-cc4f672f2e83', 'attached_at': '', 'detached_at': '', 'volume_id': '085526e0-9c7f-46ce-9f8b-d76aa54f673e', 'serial': '085526e0-9c7f-46ce-9f8b-d76aa54f673e'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 865.762659] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 865.762659] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52eae7dd-a3bb-3e4a-c9ed-7f7e7880d2c6" [ 865.762659] env[61906]: _type = "Task" [ 865.762659] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.763049] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 865.763049] env[61906]: value = "task-1356759" [ 865.763049] env[61906]: _type = "Task" [ 865.763049] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.763880] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356753, 'name': Rename_Task, 'duration_secs': 0.170939} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.767051] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 865.774434] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b286574c-47aa-43d0-a183-e14f1d198f72 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.783334] env[61906]: DEBUG oslo_vmware.api [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356758, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137244} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.791363] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 865.791667] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 865.791965] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 865.792222] env[61906]: INFO nova.compute.manager [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Took 4.78 seconds to destroy the instance on the hypervisor. [ 865.792520] env[61906]: DEBUG oslo.service.loopingcall [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.793085] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356759, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.793378] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52eae7dd-a3bb-3e4a-c9ed-7f7e7880d2c6, 'name': SearchDatastore_Task, 'duration_secs': 0.012333} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.793696] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 865.793696] env[61906]: value = "task-1356760" [ 865.793696] env[61906]: _type = "Task" [ 865.793696] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.793999] env[61906]: DEBUG nova.compute.manager [-] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 865.794151] env[61906]: DEBUG nova.network.neutron [-] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 865.795899] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.796239] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.796788] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.796788] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.796980] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.797378] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2bac64f-a7e2-43c4-aeeb-2a615d2209d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.814181] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356760, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.816041] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.816659] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 865.817763] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b629a96-25a4-4f7d-8b99-a47f1e8c7213 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.825392] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 865.825392] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52715d08-aa27-c308-3659-d8ccffeda3cc" [ 865.825392] env[61906]: _type = "Task" [ 865.825392] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.826778] env[61906]: DEBUG nova.network.neutron [-] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 865.840182] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52715d08-aa27-c308-3659-d8ccffeda3cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.942053] env[61906]: DEBUG oslo_concurrency.lockutils [req-8be7a708-c373-45fe-a075-7fd0717ce0ff req-38c4d74d-0c7a-45a5-9eb4-5e1d7922028f service nova] Releasing lock "refresh_cache-bb53f018-8546-4a00-a4be-d3c48416b1a1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.083290] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.176105] env[61906]: INFO nova.compute.manager [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Took 20.28 seconds to build instance. [ 866.276701] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356759, 'name': ReconfigVM_Task, 'duration_secs': 0.211368} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.277144] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance '943c57f9-74c6-4465-9773-3fb01e78127b' progress to 33 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 866.309156] env[61906]: DEBUG oslo_vmware.api [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356760, 'name': PowerOnVM_Task, 'duration_secs': 0.456356} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.310010] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.310237] env[61906]: INFO nova.compute.manager [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Took 9.35 seconds to spawn the instance on the hypervisor. [ 866.310430] env[61906]: DEBUG nova.compute.manager [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.311226] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ba6e31-256a-4be1-80bd-c09a728b9882 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.335463] env[61906]: DEBUG nova.network.neutron [-] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.336574] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52715d08-aa27-c308-3659-d8ccffeda3cc, 'name': SearchDatastore_Task, 'duration_secs': 0.020769} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.337500] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6315ec0-710b-466a-9182-21c3a9b24301 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.343297] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 866.343297] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]524ac97d-73ac-8175-0196-f3bbab6c7961" [ 866.343297] env[61906]: _type = "Task" [ 866.343297] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.352599] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]524ac97d-73ac-8175-0196-f3bbab6c7961, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.580497] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.677805] env[61906]: DEBUG oslo_concurrency.lockutils [None req-364f262b-e644-4013-88ff-7bef075c7a12 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "33e288d5-9065-4606-b17a-c4c90bcbd533" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.800s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.785342] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 866.785566] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 866.785691] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.785873] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 866.786035] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.786269] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 866.786500] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 866.786663] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 866.786835] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 866.787011] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 866.787220] env[61906]: DEBUG nova.virt.hardware [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.792373] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Reconfiguring VM instance instance-0000004d to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 866.792666] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-224232e4-ea96-4e15-b51b-e8a06d2b549f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.806280] env[61906]: DEBUG nova.objects.instance [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lazy-loading 'flavor' on Instance uuid 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.815313] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 866.815313] env[61906]: value = "task-1356761" [ 866.815313] env[61906]: _type = "Task" [ 866.815313] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.828216] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356761, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.830290] env[61906]: INFO nova.compute.manager [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Took 14.39 seconds to build instance. [ 866.837646] env[61906]: INFO nova.compute.manager [-] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Took 1.04 seconds to deallocate network for instance. [ 866.855374] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]524ac97d-73ac-8175-0196-f3bbab6c7961, 'name': SearchDatastore_Task, 'duration_secs': 0.01588} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.855638] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.855899] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] bb53f018-8546-4a00-a4be-d3c48416b1a1/bb53f018-8546-4a00-a4be-d3c48416b1a1.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.856516] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-072c3aed-6ef8-4e76-81d6-b8b9a40d1844 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.863658] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 866.863658] env[61906]: value = "task-1356762" [ 866.863658] env[61906]: _type = "Task" [ 866.863658] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.873135] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356762, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.082231] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.252758] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 867.253134] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 867.253210] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Deleting the datastore file [datastore1] 3a6da3bd-8e28-4c23-8660-f32f5c862d02 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.253499] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-324c4088-2e89-4a26-b1b7-c193d12dcb1f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.270890] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 867.270890] env[61906]: value = "task-1356763" [ 867.270890] env[61906]: _type = "Task" [ 867.270890] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.272835] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 867.273265] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 867.273631] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Deleting the datastore file [datastore1] 63cbf664-078b-40c5-92bb-7a5a8d8aea3b {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.278653] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d5d54d34-b9a2-4739-865b-bfa7e461dd34 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.288825] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356763, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.290871] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 867.290871] env[61906]: value = "task-1356764" [ 867.290871] env[61906]: _type = "Task" [ 867.290871] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.307959] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356764, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.312695] env[61906]: DEBUG oslo_concurrency.lockutils [None req-705d5baf-0f36-46b2-a26d-b4d837ea52e1 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.211s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.330163] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356761, 'name': ReconfigVM_Task, 'duration_secs': 0.251127} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.330335] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Reconfigured VM instance instance-0000004d to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 867.331431] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76804f0f-1b80-438c-8469-2194e0f4d8c6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.335311] env[61906]: DEBUG oslo_concurrency.lockutils [None req-58b1c0cf-e400-43cd-ac18-b46bd3288329 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.915s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.352714] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.353028] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.353276] env[61906]: DEBUG nova.objects.instance [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lazy-loading 'resources' on Instance uuid 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.362275] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 943c57f9-74c6-4465-9773-3fb01e78127b/943c57f9-74c6-4465-9773-3fb01e78127b.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.363407] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5603249b-57bd-45c9-90e9-249b9ce2211a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.389623] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356762, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.391803] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 867.391803] env[61906]: value = "task-1356765" [ 867.391803] env[61906]: _type = "Task" [ 867.391803] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.402449] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356765, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.585048] env[61906]: DEBUG oslo_vmware.api [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356740, 'name': ReconfigVM_Task, 'duration_secs': 7.177652} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.586000] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.586276] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Reconfigured VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 867.596043] env[61906]: DEBUG nova.compute.manager [req-5b71bd45-1e1e-47e8-aa02-2e0d426f1cf0 req-ec05ae46-ada1-477d-8dcd-3da9a179a9e3 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Received event network-changed-eb78ba42-034c-44d7-9dd3-4157cd702c8c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.596314] env[61906]: DEBUG nova.compute.manager [req-5b71bd45-1e1e-47e8-aa02-2e0d426f1cf0 req-ec05ae46-ada1-477d-8dcd-3da9a179a9e3 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Refreshing instance network info cache due to event network-changed-eb78ba42-034c-44d7-9dd3-4157cd702c8c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 867.596553] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b71bd45-1e1e-47e8-aa02-2e0d426f1cf0 req-ec05ae46-ada1-477d-8dcd-3da9a179a9e3 service nova] Acquiring lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.596702] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b71bd45-1e1e-47e8-aa02-2e0d426f1cf0 req-ec05ae46-ada1-477d-8dcd-3da9a179a9e3 service nova] Acquired lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.596868] env[61906]: DEBUG nova.network.neutron [req-5b71bd45-1e1e-47e8-aa02-2e0d426f1cf0 req-ec05ae46-ada1-477d-8dcd-3da9a179a9e3 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Refreshing network info cache for port eb78ba42-034c-44d7-9dd3-4157cd702c8c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.782751] env[61906]: DEBUG oslo_vmware.api [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356763, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.363522} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.783111] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.783237] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.783421] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.783603] env[61906]: INFO nova.compute.manager [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Took 7.64 seconds to destroy the instance on the hypervisor. [ 867.783851] env[61906]: DEBUG oslo.service.loopingcall [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.784084] env[61906]: DEBUG nova.compute.manager [-] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.784200] env[61906]: DEBUG nova.network.neutron [-] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 867.801266] env[61906]: DEBUG oslo_vmware.api [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356764, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.343076} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.801527] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.801742] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.801967] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.802065] env[61906]: INFO nova.compute.manager [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Took 7.72 seconds to destroy the instance on the hypervisor. [ 867.802304] env[61906]: DEBUG oslo.service.loopingcall [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.802505] env[61906]: DEBUG nova.compute.manager [-] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.802597] env[61906]: DEBUG nova.network.neutron [-] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 867.887831] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356762, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.733069} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.888372] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] bb53f018-8546-4a00-a4be-d3c48416b1a1/bb53f018-8546-4a00-a4be-d3c48416b1a1.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 867.888592] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 867.888848] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e40b29a7-1abb-4878-8082-ec4d764b7db9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.901543] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 867.901543] env[61906]: value = "task-1356766" [ 867.901543] env[61906]: _type = "Task" [ 867.901543] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.905208] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356765, 'name': ReconfigVM_Task, 'duration_secs': 0.434094} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.909822] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 943c57f9-74c6-4465-9773-3fb01e78127b/943c57f9-74c6-4465-9773-3fb01e78127b.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.910302] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance '943c57f9-74c6-4465-9773-3fb01e78127b' progress to 50 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 867.920511] env[61906]: DEBUG nova.compute.manager [req-2e292534-7325-4b83-84e9-a036008715ed req-29f7a33d-2e37-4bb2-a3d2-24dcf43eb0b4 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Received event network-vif-deleted-38287aaf-c4d5-440a-b756-6502a2c7dc02 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.922148] env[61906]: INFO nova.compute.manager [req-2e292534-7325-4b83-84e9-a036008715ed req-29f7a33d-2e37-4bb2-a3d2-24dcf43eb0b4 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Neutron deleted interface 38287aaf-c4d5-440a-b756-6502a2c7dc02; detaching it from the instance and deleting it from the info cache [ 867.922148] env[61906]: DEBUG nova.network.neutron [req-2e292534-7325-4b83-84e9-a036008715ed req-29f7a33d-2e37-4bb2-a3d2-24dcf43eb0b4 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Updating instance_info_cache with network_info: [{"id": "76876157-42ef-4519-88b2-ca9a600e3b41", "address": "fa:16:3e:66:6f:62", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76876157-42", "ovs_interfaceid": "76876157-42ef-4519-88b2-ca9a600e3b41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.929211] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.930104] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.930871] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356766, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.075528] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "33e288d5-9065-4606-b17a-c4c90bcbd533" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.075645] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "33e288d5-9065-4606-b17a-c4c90bcbd533" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.075864] env[61906]: INFO nova.compute.manager [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Shelving [ 868.135552] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e58355-a73c-424b-a9ad-4d57fd01d291 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.150096] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45658f7b-e38d-4ea8-befc-54a91c615e61 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.189486] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd382f44-a354-472d-9b97-7f8e6babd464 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.198583] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d069c9b-1b62-4496-8f1b-55f5fb3e6ed6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.213245] env[61906]: DEBUG nova.compute.provider_tree [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.414970] env[61906]: DEBUG nova.network.neutron [req-5b71bd45-1e1e-47e8-aa02-2e0d426f1cf0 req-ec05ae46-ada1-477d-8dcd-3da9a179a9e3 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updated VIF entry in instance network info cache for port eb78ba42-034c-44d7-9dd3-4157cd702c8c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 868.415425] env[61906]: DEBUG nova.network.neutron [req-5b71bd45-1e1e-47e8-aa02-2e0d426f1cf0 req-ec05ae46-ada1-477d-8dcd-3da9a179a9e3 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updating instance_info_cache with network_info: [{"id": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "address": "fa:16:3e:f0:1b:68", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb78ba42-03", "ovs_interfaceid": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.424148] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7074172-59fb-4ac9-b261-9559420ebd03 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.435180] env[61906]: DEBUG oslo_concurrency.lockutils [req-2e292534-7325-4b83-84e9-a036008715ed req-29f7a33d-2e37-4bb2-a3d2-24dcf43eb0b4 service nova] Acquiring lock "5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.435180] env[61906]: DEBUG oslo_concurrency.lockutils [req-2e292534-7325-4b83-84e9-a036008715ed req-29f7a33d-2e37-4bb2-a3d2-24dcf43eb0b4 service nova] Acquired lock "5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.435180] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356766, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094788} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.435852] env[61906]: INFO nova.compute.manager [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Detaching volume 085526e0-9c7f-46ce-9f8b-d76aa54f673e [ 868.438516] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8880ac57-fe15-451f-945b-d2dc64620617 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.441622] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 868.458353] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2193adc1-5cf3-4314-8c5b-0b8b1b24cead {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.465323] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05392388-5c60-4269-b359-b1f439725629 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.481950] env[61906]: DEBUG oslo_concurrency.lockutils [req-2e292534-7325-4b83-84e9-a036008715ed req-29f7a33d-2e37-4bb2-a3d2-24dcf43eb0b4 service nova] Releasing lock "5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.482280] env[61906]: WARNING nova.compute.manager [req-2e292534-7325-4b83-84e9-a036008715ed req-29f7a33d-2e37-4bb2-a3d2-24dcf43eb0b4 service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Detach interface failed, port_id=38287aaf-c4d5-440a-b756-6502a2c7dc02, reason: No device with interface-id 38287aaf-c4d5-440a-b756-6502a2c7dc02 exists on VM: nova.exception.NotFound: No device with interface-id 38287aaf-c4d5-440a-b756-6502a2c7dc02 exists on VM [ 868.503826] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] bb53f018-8546-4a00-a4be-d3c48416b1a1/bb53f018-8546-4a00-a4be-d3c48416b1a1.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 868.518754] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc140836-7ba2-40b0-bb2f-a465e457b9ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.536998] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance '943c57f9-74c6-4465-9773-3fb01e78127b' progress to 67 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 868.541543] env[61906]: DEBUG nova.network.neutron [-] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.544707] env[61906]: INFO nova.virt.block_device [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Attempting to driver detach volume 085526e0-9c7f-46ce-9f8b-d76aa54f673e from mountpoint /dev/sdb [ 868.544707] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 868.544707] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288996', 'volume_id': '085526e0-9c7f-46ce-9f8b-d76aa54f673e', 'name': 'volume-085526e0-9c7f-46ce-9f8b-d76aa54f673e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '51fa46a9-b1b3-4115-b1d7-cc4f672f2e83', 'attached_at': '', 'detached_at': '', 'volume_id': '085526e0-9c7f-46ce-9f8b-d76aa54f673e', 'serial': '085526e0-9c7f-46ce-9f8b-d76aa54f673e'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 868.544707] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6eab6a4-5fa0-45d0-ac6b-93b44d3ab156 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.571390] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cafbfe2-4ab0-4c12-9e4f-856174c2566c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.574293] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 868.574293] env[61906]: value = "task-1356767" [ 868.574293] env[61906]: _type = "Task" [ 868.574293] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.585012] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5977c00e-9435-41a3-b85a-360c98dc77f9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.591847] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.592431] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356767, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.592633] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f288a9c1-0669-45e2-8b1c-12df6eb017b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.615091] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4290ea25-ca5c-4bbb-8c0f-0060ba0f5aaf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.618026] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 868.618026] env[61906]: value = "task-1356768" [ 868.618026] env[61906]: _type = "Task" [ 868.618026] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.632169] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] The volume has not been displaced from its original location: [datastore2] volume-085526e0-9c7f-46ce-9f8b-d76aa54f673e/volume-085526e0-9c7f-46ce-9f8b-d76aa54f673e.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 868.637283] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Reconfiguring VM instance instance-00000043 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 868.638051] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-039ff421-a9fd-4423-a701-f8d4ef42f02e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.655046] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356768, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.661787] env[61906]: DEBUG oslo_vmware.api [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 868.661787] env[61906]: value = "task-1356769" [ 868.661787] env[61906]: _type = "Task" [ 868.661787] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.669614] env[61906]: DEBUG oslo_vmware.api [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356769, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.716941] env[61906]: DEBUG nova.scheduler.client.report [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.880062] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "5adaa660-b736-4c11-9141-846cf475ccd5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.880370] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "5adaa660-b736-4c11-9141-846cf475ccd5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.880592] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "5adaa660-b736-4c11-9141-846cf475ccd5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.880787] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "5adaa660-b736-4c11-9141-846cf475ccd5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.880969] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "5adaa660-b736-4c11-9141-846cf475ccd5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.883344] env[61906]: INFO nova.compute.manager [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Terminating instance [ 868.885392] env[61906]: DEBUG nova.compute.manager [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 868.885601] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 868.886494] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd362fd0-a658-4b98-a97e-a84e81ee1a50 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.895033] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.895220] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.895401] env[61906]: DEBUG nova.network.neutron [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.896634] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.896882] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1a2f8ac-506f-454a-a8d1-a7406e27c3b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.905071] env[61906]: DEBUG oslo_vmware.api [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 868.905071] env[61906]: value = "task-1356770" [ 868.905071] env[61906]: _type = "Task" [ 868.905071] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.915687] env[61906]: DEBUG oslo_vmware.api [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356770, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.918437] env[61906]: DEBUG oslo_concurrency.lockutils [req-5b71bd45-1e1e-47e8-aa02-2e0d426f1cf0 req-ec05ae46-ada1-477d-8dcd-3da9a179a9e3 service nova] Releasing lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.994716] env[61906]: DEBUG nova.network.neutron [-] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.048866] env[61906]: INFO nova.compute.manager [-] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Took 1.25 seconds to deallocate network for instance. [ 869.084874] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356767, 'name': ReconfigVM_Task, 'duration_secs': 0.29912} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.085771] env[61906]: DEBUG nova.network.neutron [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Port ba127004-dc19-4fc4-b84b-97584ac68f34 binding to destination host cpu-1 is already ACTIVE {{(pid=61906) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 869.087282] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Reconfigured VM instance instance-00000052 to attach disk [datastore2] bb53f018-8546-4a00-a4be-d3c48416b1a1/bb53f018-8546-4a00-a4be-d3c48416b1a1.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.088641] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7138d1af-da11-47d8-8a95-db035ed3ff3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.097449] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 869.097449] env[61906]: value = "task-1356771" [ 869.097449] env[61906]: _type = "Task" [ 869.097449] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.108948] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356771, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.130065] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356768, 'name': PowerOffVM_Task, 'duration_secs': 0.178998} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.130065] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.130769] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37e5fd9-4732-42ac-97ca-7d6707262f86 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.151312] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac8e01c3-ad98-45f8-bcf5-66f58eabafea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.170940] env[61906]: DEBUG oslo_vmware.api [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356769, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.222193] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.869s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.249185] env[61906]: INFO nova.scheduler.client.report [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Deleted allocations for instance 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44 [ 869.415345] env[61906]: DEBUG oslo_vmware.api [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356770, 'name': PowerOffVM_Task, 'duration_secs': 0.17396} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.415671] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.415792] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 869.416054] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb81bdb5-c230-43d0-b0a0-ffaa4123b823 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.494597] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 869.494901] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 869.495128] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Deleting the datastore file [datastore2] 5adaa660-b736-4c11-9141-846cf475ccd5 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.495661] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf418c65-a392-4194-bcb7-b454230eab0c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.498752] env[61906]: INFO nova.compute.manager [-] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Took 1.71 seconds to deallocate network for instance. [ 869.507024] env[61906]: DEBUG oslo_vmware.api [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 869.507024] env[61906]: value = "task-1356773" [ 869.507024] env[61906]: _type = "Task" [ 869.507024] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.513999] env[61906]: DEBUG oslo_vmware.api [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356773, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.557396] env[61906]: DEBUG oslo_concurrency.lockutils [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.557671] env[61906]: DEBUG oslo_concurrency.lockutils [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.557899] env[61906]: DEBUG nova.objects.instance [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lazy-loading 'resources' on Instance uuid 63cbf664-078b-40c5-92bb-7a5a8d8aea3b {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.615186] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356771, 'name': Rename_Task, 'duration_secs': 0.150026} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.620019] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.620019] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18c03d32-3be4-4aeb-8215-bee19313d1b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.627509] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 869.627509] env[61906]: value = "task-1356774" [ 869.627509] env[61906]: _type = "Task" [ 869.627509] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.637754] env[61906]: DEBUG nova.compute.manager [req-a49b87f6-2dfc-490f-b6d8-1fbece11fa7d req-0f0f02a7-13f7-4fbb-a604-fed34c328324 service nova] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Received event network-vif-deleted-3d71c788-b4a5-4f7a-8184-d9bbe379bd70 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.637964] env[61906]: DEBUG nova.compute.manager [req-a49b87f6-2dfc-490f-b6d8-1fbece11fa7d req-0f0f02a7-13f7-4fbb-a604-fed34c328324 service nova] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Received event network-vif-deleted-ba141737-69f3-4c17-a4af-b5c5a4447d6b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.645301] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356774, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.662137] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 869.662827] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-981d99d0-8046-49b9-a59b-598799e2bd69 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.674084] env[61906]: DEBUG oslo_vmware.api [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356769, 'name': ReconfigVM_Task, 'duration_secs': 0.657635} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.678023] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Reconfigured VM instance instance-00000043 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 869.683615] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 869.683615] env[61906]: value = "task-1356775" [ 869.683615] env[61906]: _type = "Task" [ 869.683615] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.684226] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1235cd09-be63-48f4-8478-b05fa34f79a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.703726] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356775, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.705259] env[61906]: DEBUG oslo_vmware.api [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 869.705259] env[61906]: value = "task-1356776" [ 869.705259] env[61906]: _type = "Task" [ 869.705259] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.713690] env[61906]: DEBUG oslo_vmware.api [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356776, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.731216] env[61906]: DEBUG nova.network.neutron [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Updating instance_info_cache with network_info: [{"id": "76876157-42ef-4519-88b2-ca9a600e3b41", "address": "fa:16:3e:66:6f:62", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76876157-42", "ovs_interfaceid": "76876157-42ef-4519-88b2-ca9a600e3b41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.759123] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a6307dd5-9e90-409a-9ebb-d8fc3b4750a7 tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "31d0f640-47f4-4b72-b9fe-6ef4c76c6b44" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.879s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.932175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "a5f780a2-0cb1-4da8-8276-82b9653bf6a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.932661] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "a5f780a2-0cb1-4da8-8276-82b9653bf6a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.933074] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "a5f780a2-0cb1-4da8-8276-82b9653bf6a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.933407] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "a5f780a2-0cb1-4da8-8276-82b9653bf6a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.933670] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "a5f780a2-0cb1-4da8-8276-82b9653bf6a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.936358] env[61906]: INFO nova.compute.manager [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Terminating instance [ 869.938283] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "refresh_cache-a5f780a2-0cb1-4da8-8276-82b9653bf6a7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.938471] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquired lock "refresh_cache-a5f780a2-0cb1-4da8-8276-82b9653bf6a7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.938692] env[61906]: DEBUG nova.network.neutron [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 870.006931] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.021249] env[61906]: DEBUG oslo_vmware.api [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356773, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141131} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.021542] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.021737] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 870.021919] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 870.022118] env[61906]: INFO nova.compute.manager [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Took 1.14 seconds to destroy the instance on the hypervisor. [ 870.022363] env[61906]: DEBUG oslo.service.loopingcall [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.022565] env[61906]: DEBUG nova.compute.manager [-] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 870.022658] env[61906]: DEBUG nova.network.neutron [-] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 870.115118] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "943c57f9-74c6-4465-9773-3fb01e78127b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.115383] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "943c57f9-74c6-4465-9773-3fb01e78127b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.115570] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "943c57f9-74c6-4465-9773-3fb01e78127b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.140890] env[61906]: DEBUG oslo_vmware.api [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356774, 'name': PowerOnVM_Task, 'duration_secs': 0.486674} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.143695] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 870.143919] env[61906]: INFO nova.compute.manager [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Took 8.16 seconds to spawn the instance on the hypervisor. [ 870.144123] env[61906]: DEBUG nova.compute.manager [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.145148] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964e0928-df7e-4ab7-8010-36ec7488c05e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.208314] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356775, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.220205] env[61906]: DEBUG oslo_vmware.api [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356776, 'name': ReconfigVM_Task, 'duration_secs': 0.157626} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.220539] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-288996', 'volume_id': '085526e0-9c7f-46ce-9f8b-d76aa54f673e', 'name': 'volume-085526e0-9c7f-46ce-9f8b-d76aa54f673e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '51fa46a9-b1b3-4115-b1d7-cc4f672f2e83', 'attached_at': '', 'detached_at': '', 'volume_id': '085526e0-9c7f-46ce-9f8b-d76aa54f673e', 'serial': '085526e0-9c7f-46ce-9f8b-d76aa54f673e'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 870.233892] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "refresh_cache-5adaa660-b736-4c11-9141-846cf475ccd5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.310216] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d10229-5043-4e5a-a5b3-5e677037493d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.320371] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5830d19a-7b76-4312-8f5d-eede1ab3fec5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.351601] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c403649c-d225-420e-82e3-36b780289ade {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.359881] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-417e8c54-74b5-447d-95d7-762cb1f5dd17 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.373490] env[61906]: DEBUG nova.compute.provider_tree [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.459949] env[61906]: DEBUG nova.network.neutron [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 870.542508] env[61906]: DEBUG nova.network.neutron [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.673955] env[61906]: INFO nova.compute.manager [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Took 14.61 seconds to build instance. [ 870.709845] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356775, 'name': CreateSnapshot_Task, 'duration_secs': 0.660779} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.711345] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 870.712711] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b79b67-3cab-4953-8f55-96db3fb94225 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.738390] env[61906]: DEBUG oslo_concurrency.lockutils [None req-39b7b9b9-b669-4ae0-8a8b-2f686a1bb90a tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-5adaa660-b736-4c11-9141-846cf475ccd5-38287aaf-c4d5-440a-b756-6502a2c7dc02" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.994s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.784729] env[61906]: DEBUG nova.objects.instance [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lazy-loading 'flavor' on Instance uuid 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.876759] env[61906]: DEBUG nova.scheduler.client.report [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.049117] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Releasing lock "refresh_cache-a5f780a2-0cb1-4da8-8276-82b9653bf6a7" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.049581] env[61906]: DEBUG nova.compute.manager [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 871.049783] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 871.050836] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e63630a-ca60-4b82-a3ec-9b6da2cc0708 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.060172] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 871.060172] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e259fa3d-39b2-409e-96a4-38db22eee264 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.067115] env[61906]: DEBUG oslo_vmware.api [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 871.067115] env[61906]: value = "task-1356777" [ 871.067115] env[61906]: _type = "Task" [ 871.067115] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.076779] env[61906]: DEBUG oslo_vmware.api [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356777, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.174997] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.175227] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.175407] env[61906]: DEBUG nova.network.neutron [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.176775] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f031e25c-fa92-4eab-ad6f-4f47bb074ea8 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "bb53f018-8546-4a00-a4be-d3c48416b1a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.120s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.226091] env[61906]: DEBUG nova.network.neutron [-] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.236107] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 871.236772] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ffac8fb4-23ef-433b-aae0-923d353a9a3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.248117] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 871.248117] env[61906]: value = "task-1356778" [ 871.248117] env[61906]: _type = "Task" [ 871.248117] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.257206] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356778, 'name': CloneVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.383786] env[61906]: DEBUG oslo_concurrency.lockutils [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.826s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.388428] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.381s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.388509] env[61906]: DEBUG nova.objects.instance [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lazy-loading 'resources' on Instance uuid 3a6da3bd-8e28-4c23-8660-f32f5c862d02 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.409585] env[61906]: INFO nova.scheduler.client.report [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Deleted allocations for instance 63cbf664-078b-40c5-92bb-7a5a8d8aea3b [ 871.580034] env[61906]: DEBUG oslo_vmware.api [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356777, 'name': PowerOffVM_Task, 'duration_secs': 0.123472} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.580034] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 871.580460] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 871.580558] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1d51ea3a-c31e-4d3f-be5b-141a5916312d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.610997] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 871.611235] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 871.611422] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Deleting the datastore file [datastore2] a5f780a2-0cb1-4da8-8276-82b9653bf6a7 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 871.611710] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-593d0211-fb1b-4e25-9855-98e5061c173e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.618121] env[61906]: DEBUG oslo_vmware.api [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for the task: (returnval){ [ 871.618121] env[61906]: value = "task-1356780" [ 871.618121] env[61906]: _type = "Task" [ 871.618121] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.626156] env[61906]: DEBUG oslo_vmware.api [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356780, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.655195] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ed4fa195-3858-4452-846e-087995a57e41 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "e133605d-c630-4b9e-a314-bf496c853710" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.655195] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ed4fa195-3858-4452-846e-087995a57e41 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.655337] env[61906]: DEBUG nova.compute.manager [None req-ed4fa195-3858-4452-846e-087995a57e41 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 871.656254] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79315ca7-afdd-45d4-a4cc-78837d62c8b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.661497] env[61906]: DEBUG nova.compute.manager [req-88523c47-f9df-4532-8d9d-ee293682e5e5 req-030bbbe2-bcbe-47a6-9b17-98c7ea5088eb service nova] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Received event network-vif-deleted-76876157-42ef-4519-88b2-ca9a600e3b41 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.667036] env[61906]: DEBUG nova.compute.manager [None req-ed4fa195-3858-4452-846e-087995a57e41 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61906) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 871.667128] env[61906]: DEBUG nova.objects.instance [None req-ed4fa195-3858-4452-846e-087995a57e41 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lazy-loading 'flavor' on Instance uuid e133605d-c630-4b9e-a314-bf496c853710 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.737898] env[61906]: INFO nova.compute.manager [-] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Took 1.72 seconds to deallocate network for instance. [ 871.758852] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356778, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.792484] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e29dbd37-2c09-4ae4-8d55-476f410f56ed tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.863s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.886139] env[61906]: DEBUG nova.network.neutron [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance_info_cache with network_info: [{"id": "ba127004-dc19-4fc4-b84b-97584ac68f34", "address": "fa:16:3e:d1:ad:74", "network": {"id": "9e0705ba-7b9f-46fb-8603-01598a888212", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-705593104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb85e88fd4f54e0db2ff131f81137f64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba127004-dc", "ovs_interfaceid": "ba127004-dc19-4fc4-b84b-97584ac68f34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.917551] env[61906]: DEBUG oslo_concurrency.lockutils [None req-781cd904-7500-4c3c-876b-758617bec868 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "63cbf664-078b-40c5-92bb-7a5a8d8aea3b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.847s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.009025] env[61906]: DEBUG nova.compute.manager [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.011974] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898f5ec4-30f8-46d5-b7e1-cab234e8f971 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.057896] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-400b7253-0084-4a0d-8f47-0df492a6601e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.064446] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436d0c53-0c2c-48e5-8f7d-7a114cdd3fd1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.095442] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b209f982-e979-4f18-b84e-59226681f7b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.102974] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61bedee9-3a1f-4811-becb-daa8f7cca178 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.119091] env[61906]: DEBUG nova.compute.provider_tree [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.129521] env[61906]: DEBUG oslo_vmware.api [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Task: {'id': task-1356780, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.352055} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.129789] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 872.129974] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 872.130173] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 872.130350] env[61906]: INFO nova.compute.manager [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Took 1.08 seconds to destroy the instance on the hypervisor. [ 872.130597] env[61906]: DEBUG oslo.service.loopingcall [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.131022] env[61906]: DEBUG nova.compute.manager [-] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 872.131122] env[61906]: DEBUG nova.network.neutron [-] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 872.146935] env[61906]: DEBUG nova.network.neutron [-] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.171151] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed4fa195-3858-4452-846e-087995a57e41 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 872.171429] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d5002aa9-add1-4eaf-b7b5-740bd27608c6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.178914] env[61906]: DEBUG oslo_vmware.api [None req-ed4fa195-3858-4452-846e-087995a57e41 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 872.178914] env[61906]: value = "task-1356781" [ 872.178914] env[61906]: _type = "Task" [ 872.178914] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.188352] env[61906]: DEBUG oslo_vmware.api [None req-ed4fa195-3858-4452-846e-087995a57e41 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356781, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.244141] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.260833] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356778, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.391965] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.396377] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.396722] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.397051] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.397379] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.397616] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.400796] env[61906]: INFO nova.compute.manager [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Terminating instance [ 872.402794] env[61906]: DEBUG nova.compute.manager [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 872.403014] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 872.404068] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eeb9dcc-63a4-45f4-bdee-44aa8e3b7a4b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.412664] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 872.412922] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d512148-490c-477d-a444-84002de0998b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.419987] env[61906]: DEBUG oslo_vmware.api [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 872.419987] env[61906]: value = "task-1356782" [ 872.419987] env[61906]: _type = "Task" [ 872.419987] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.427527] env[61906]: DEBUG oslo_vmware.api [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356782, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.522536] env[61906]: INFO nova.compute.manager [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] instance snapshotting [ 872.525440] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0bf804-8633-4fc1-968c-738f43278a33 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.546767] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32c51db-9b62-41f8-9752-5905361260ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.625287] env[61906]: DEBUG nova.scheduler.client.report [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.650027] env[61906]: DEBUG nova.network.neutron [-] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.689127] env[61906]: DEBUG oslo_vmware.api [None req-ed4fa195-3858-4452-846e-087995a57e41 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356781, 'name': PowerOffVM_Task, 'duration_secs': 0.237908} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.689443] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed4fa195-3858-4452-846e-087995a57e41 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 872.689654] env[61906]: DEBUG nova.compute.manager [None req-ed4fa195-3858-4452-846e-087995a57e41 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.690498] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05164e38-3a04-43d2-800e-7b407add10ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.760838] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356778, 'name': CloneVM_Task, 'duration_secs': 1.47846} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.761237] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Created linked-clone VM from snapshot [ 872.762046] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7efb6090-8ed7-42fe-b686-b7816e585956 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.769624] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Uploading image 629c5099-4091-4b95-8c5a-a3a2b13b491d {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 872.791810] env[61906]: DEBUG oslo_vmware.rw_handles [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 872.791810] env[61906]: value = "vm-289000" [ 872.791810] env[61906]: _type = "VirtualMachine" [ 872.791810] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 872.792105] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5823839c-b637-4ff9-9bda-4dcac95ad3a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.799682] env[61906]: DEBUG oslo_vmware.rw_handles [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lease: (returnval){ [ 872.799682] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c628f4-8a09-cba5-4995-b55cbd44f4c2" [ 872.799682] env[61906]: _type = "HttpNfcLease" [ 872.799682] env[61906]: } obtained for exporting VM: (result){ [ 872.799682] env[61906]: value = "vm-289000" [ 872.799682] env[61906]: _type = "VirtualMachine" [ 872.799682] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 872.799985] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the lease: (returnval){ [ 872.799985] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c628f4-8a09-cba5-4995-b55cbd44f4c2" [ 872.799985] env[61906]: _type = "HttpNfcLease" [ 872.799985] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 872.806702] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 872.806702] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c628f4-8a09-cba5-4995-b55cbd44f4c2" [ 872.806702] env[61906]: _type = "HttpNfcLease" [ 872.806702] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 872.919637] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e127c8b1-77e3-4a09-bd2f-da2f4adc3d4d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.931232] env[61906]: DEBUG oslo_vmware.api [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356782, 'name': PowerOffVM_Task, 'duration_secs': 0.291001} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.944426] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 872.944606] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.945011] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da8d403b-530e-4efe-89c0-f56f7dd9f021 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.947176] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46d53254-e13e-4e01-bdc9-b6ff49fb0506 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.956068] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance '943c57f9-74c6-4465-9773-3fb01e78127b' progress to 83 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 873.058807] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 873.059175] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a0b322c2-916b-4891-9937-7f1b744c8fe8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.066967] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 873.066967] env[61906]: value = "task-1356785" [ 873.066967] env[61906]: _type = "Task" [ 873.066967] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.076977] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356785, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.081836] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.082362] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.082362] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Deleting the datastore file [datastore2] 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.082676] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6974781d-7d92-40ed-b915-6077dd4bcc4e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.089010] env[61906]: DEBUG oslo_vmware.api [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 873.089010] env[61906]: value = "task-1356786" [ 873.089010] env[61906]: _type = "Task" [ 873.089010] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.098236] env[61906]: DEBUG oslo_vmware.api [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356786, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.130698] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.743s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.133055] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.889s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.133324] env[61906]: DEBUG nova.objects.instance [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'resources' on Instance uuid 5adaa660-b736-4c11-9141-846cf475ccd5 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.149803] env[61906]: INFO nova.scheduler.client.report [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Deleted allocations for instance 3a6da3bd-8e28-4c23-8660-f32f5c862d02 [ 873.154061] env[61906]: INFO nova.compute.manager [-] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Took 1.02 seconds to deallocate network for instance. [ 873.204860] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ed4fa195-3858-4452-846e-087995a57e41 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.549s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.308496] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 873.308496] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c628f4-8a09-cba5-4995-b55cbd44f4c2" [ 873.308496] env[61906]: _type = "HttpNfcLease" [ 873.308496] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 873.308878] env[61906]: DEBUG oslo_vmware.rw_handles [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 873.308878] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c628f4-8a09-cba5-4995-b55cbd44f4c2" [ 873.308878] env[61906]: _type = "HttpNfcLease" [ 873.308878] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 873.309643] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5803a83c-9aae-4718-9b47-93f05dfc4223 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.317254] env[61906]: DEBUG oslo_vmware.rw_handles [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529482d4-9405-b814-0797-c1922ece743b/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 873.317646] env[61906]: DEBUG oslo_vmware.rw_handles [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529482d4-9405-b814-0797-c1922ece743b/disk-0.vmdk for reading. {{(pid=61906) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 873.406807] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6d92d3ad-2f31-4a15-8b72-f2e814513bdd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.464469] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 873.464802] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05230699-61cc-4d1c-a88f-3462dc70b1aa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.476826] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 873.476826] env[61906]: value = "task-1356787" [ 873.476826] env[61906]: _type = "Task" [ 873.476826] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.486975] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356787, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.577158] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356785, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.598687] env[61906]: DEBUG oslo_vmware.api [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356786, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134013} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.599035] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.599284] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.599462] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.599643] env[61906]: INFO nova.compute.manager [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Took 1.20 seconds to destroy the instance on the hypervisor. [ 873.599999] env[61906]: DEBUG oslo.service.loopingcall [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.600231] env[61906]: DEBUG nova.compute.manager [-] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.600324] env[61906]: DEBUG nova.network.neutron [-] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 873.660903] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.661487] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f1154711-ad7b-49bb-8b15-7219a5413f6b tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "3a6da3bd-8e28-4c23-8660-f32f5c862d02" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.525s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.837043] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f73723f-c15e-46ba-9c9b-26a5131ec34e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.847173] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4ec193-de20-48a5-b0fd-adfcea914b44 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.884629] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c69166-a65d-4dde-a7eb-9b290088389b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.893588] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70104ff3-0f7d-49ec-9c79-e6feb7759443 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.910057] env[61906]: DEBUG nova.compute.provider_tree [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.990841] env[61906]: DEBUG oslo_vmware.api [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356787, 'name': PowerOnVM_Task, 'duration_secs': 0.483211} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.991476] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 873.991878] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8998f6-dd12-4db9-931e-3894d1bf08ad tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance '943c57f9-74c6-4465-9773-3fb01e78127b' progress to 100 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 874.027083] env[61906]: DEBUG nova.objects.instance [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lazy-loading 'flavor' on Instance uuid e133605d-c630-4b9e-a314-bf496c853710 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.079168] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356785, 'name': CreateSnapshot_Task, 'duration_secs': 0.811536} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.079852] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 874.081116] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e3f545-d600-4350-a401-5738182913ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.141857] env[61906]: DEBUG nova.compute.manager [req-8b5bccbc-708e-4a00-8ad6-3185236380b1 req-72961dce-0fb0-44db-86d5-c00b74bb22eb service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Received event network-vif-deleted-a4af9140-6653-413f-a40e-0b0c021dc68c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.141857] env[61906]: INFO nova.compute.manager [req-8b5bccbc-708e-4a00-8ad6-3185236380b1 req-72961dce-0fb0-44db-86d5-c00b74bb22eb service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Neutron deleted interface a4af9140-6653-413f-a40e-0b0c021dc68c; detaching it from the instance and deleting it from the info cache [ 874.141857] env[61906]: DEBUG nova.network.neutron [req-8b5bccbc-708e-4a00-8ad6-3185236380b1 req-72961dce-0fb0-44db-86d5-c00b74bb22eb service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.413917] env[61906]: DEBUG nova.scheduler.client.report [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.531700] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.532100] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.532738] env[61906]: DEBUG nova.network.neutron [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 874.532738] env[61906]: DEBUG nova.objects.instance [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lazy-loading 'info_cache' on Instance uuid e133605d-c630-4b9e-a314-bf496c853710 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.604964] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 874.605507] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e8dbf201-adf1-4353-aef7-5aab2d768aa0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.616151] env[61906]: DEBUG nova.network.neutron [-] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.618182] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 874.618182] env[61906]: value = "task-1356788" [ 874.618182] env[61906]: _type = "Task" [ 874.618182] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.627509] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356788, 'name': CloneVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.644068] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-19d2e3f8-6c14-4dde-970a-745437797241 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.653609] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f489c2-0ee6-4022-bbe7-ed61d182c099 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.684942] env[61906]: DEBUG nova.compute.manager [req-8b5bccbc-708e-4a00-8ad6-3185236380b1 req-72961dce-0fb0-44db-86d5-c00b74bb22eb service nova] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Detach interface failed, port_id=a4af9140-6653-413f-a40e-0b0c021dc68c, reason: Instance 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 874.919610] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.922568] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.262s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.922691] env[61906]: DEBUG nova.objects.instance [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lazy-loading 'resources' on Instance uuid a5f780a2-0cb1-4da8-8276-82b9653bf6a7 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.940030] env[61906]: INFO nova.scheduler.client.report [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Deleted allocations for instance 5adaa660-b736-4c11-9141-846cf475ccd5 [ 875.036288] env[61906]: DEBUG nova.objects.base [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 875.119523] env[61906]: INFO nova.compute.manager [-] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Took 1.52 seconds to deallocate network for instance. [ 875.134273] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356788, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.371560] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "d0792521-4637-4ce7-a579-71caa80ba38c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.371892] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "d0792521-4637-4ce7-a579-71caa80ba38c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.409830] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "03bd64e4-16fc-4659-a428-a2d9e7205b81" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.410166] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "03bd64e4-16fc-4659-a428-a2d9e7205b81" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.453060] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bfeb5317-0d90-48d2-a800-c643ed1cccf5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "5adaa660-b736-4c11-9141-846cf475ccd5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.572s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.631557] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.632321] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356788, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.646934] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03365fa1-1210-499e-b9ed-53f2703ee411 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.655300] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7c5737-3a13-4a36-b2f1-1d2f60242a16 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.691393] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded2f929-4177-44e7-9770-cb8faba4875f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.697312] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f264dd2-389b-47f8-b64e-d75c155a17c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.712624] env[61906]: DEBUG nova.compute.provider_tree [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.874694] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 875.913258] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 875.935133] env[61906]: DEBUG nova.network.neutron [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance_info_cache with network_info: [{"id": "d803c8e8-7145-412c-aa3e-2545666742eb", "address": "fa:16:3e:46:59:77", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd803c8e8-71", "ovs_interfaceid": "d803c8e8-7145-412c-aa3e-2545666742eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.098588] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "943c57f9-74c6-4465-9773-3fb01e78127b" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.099709] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "943c57f9-74c6-4465-9773-3fb01e78127b" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.099709] env[61906]: DEBUG nova.compute.manager [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Going to confirm migration 1 {{(pid=61906) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 876.133603] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356788, 'name': CloneVM_Task} progress is 95%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.216327] env[61906]: DEBUG nova.scheduler.client.report [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.397886] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.433735] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.438582] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.631849] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356788, 'name': CloneVM_Task, 'duration_secs': 1.824305} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.632146] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Created linked-clone VM from snapshot [ 876.632888] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b86beab-af4b-4855-a070-93abd1c27eed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.640618] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Uploading image cc54d929-47af-46bb-8c53-a38fb410a7cc {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 876.656539] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.656735] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.656915] env[61906]: DEBUG nova.network.neutron [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 876.657117] env[61906]: DEBUG nova.objects.instance [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lazy-loading 'info_cache' on Instance uuid 943c57f9-74c6-4465-9773-3fb01e78127b {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.663759] env[61906]: DEBUG oslo_vmware.rw_handles [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 876.663759] env[61906]: value = "vm-289002" [ 876.663759] env[61906]: _type = "VirtualMachine" [ 876.663759] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 876.665212] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e9a9ac39-0797-4c7a-bab3-4662bf472be4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.672252] env[61906]: DEBUG oslo_vmware.rw_handles [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lease: (returnval){ [ 876.672252] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ef4692-473a-5869-d760-4ca8f26f751c" [ 876.672252] env[61906]: _type = "HttpNfcLease" [ 876.672252] env[61906]: } obtained for exporting VM: (result){ [ 876.672252] env[61906]: value = "vm-289002" [ 876.672252] env[61906]: _type = "VirtualMachine" [ 876.672252] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 876.672638] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the lease: (returnval){ [ 876.672638] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ef4692-473a-5869-d760-4ca8f26f751c" [ 876.672638] env[61906]: _type = "HttpNfcLease" [ 876.672638] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 876.682306] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 876.682306] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ef4692-473a-5869-d760-4ca8f26f751c" [ 876.682306] env[61906]: _type = "HttpNfcLease" [ 876.682306] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 876.722273] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.800s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.724745] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.093s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.725187] env[61906]: DEBUG nova.objects.instance [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lazy-loading 'resources' on Instance uuid 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.742473] env[61906]: INFO nova.scheduler.client.report [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Deleted allocations for instance a5f780a2-0cb1-4da8-8276-82b9653bf6a7 [ 876.941683] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.942014] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77895c14-131d-4570-8b40-34c3e7196973 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.948953] env[61906]: DEBUG oslo_vmware.api [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 876.948953] env[61906]: value = "task-1356790" [ 876.948953] env[61906]: _type = "Task" [ 876.948953] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.956562] env[61906]: DEBUG oslo_vmware.api [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356790, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.180356] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 877.180356] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ef4692-473a-5869-d760-4ca8f26f751c" [ 877.180356] env[61906]: _type = "HttpNfcLease" [ 877.180356] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 877.180714] env[61906]: DEBUG oslo_vmware.rw_handles [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 877.180714] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ef4692-473a-5869-d760-4ca8f26f751c" [ 877.180714] env[61906]: _type = "HttpNfcLease" [ 877.180714] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 877.181529] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e24e0d1f-39f8-4d29-bb93-688d8e36de10 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.189407] env[61906]: DEBUG oslo_vmware.rw_handles [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a63af9-3a7a-0cce-df05-b66591c7af0f/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 877.189607] env[61906]: DEBUG oslo_vmware.rw_handles [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a63af9-3a7a-0cce-df05-b66591c7af0f/disk-0.vmdk for reading. {{(pid=61906) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 877.259133] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e6217c28-6e9a-4034-b42c-6c38caf3937c tempest-ServerShowV247Test-1818656732 tempest-ServerShowV247Test-1818656732-project-member] Lock "a5f780a2-0cb1-4da8-8276-82b9653bf6a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.326s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.303232] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3ebf3598-5df1-47b1-a9da-bdd84ec7fc74 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.409830] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "99e5c55d-79ae-47a6-8500-79fc68291650" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.410118] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "99e5c55d-79ae-47a6-8500-79fc68291650" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.460679] env[61906]: DEBUG oslo_vmware.api [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356790, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.495821] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0791d5-276e-4ef9-9354-4bb9b552e662 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.503457] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719d4191-204c-4ae1-9fad-18ccb88a979e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.534910] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f707f2a7-4166-4f07-bce3-c345d24bda34 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.543250] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97cf01cb-a6ee-4660-b239-a75fbd23af7e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.560276] env[61906]: DEBUG nova.compute.provider_tree [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.916017] env[61906]: DEBUG nova.compute.manager [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 877.964750] env[61906]: DEBUG oslo_vmware.api [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356790, 'name': PowerOnVM_Task, 'duration_secs': 0.550244} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.964750] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.964750] env[61906]: DEBUG nova.compute.manager [None req-6c73cf70-803e-4b83-a8b0-82b0daf4aff4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.964750] env[61906]: DEBUG nova.network.neutron [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance_info_cache with network_info: [{"id": "ba127004-dc19-4fc4-b84b-97584ac68f34", "address": "fa:16:3e:d1:ad:74", "network": {"id": "9e0705ba-7b9f-46fb-8603-01598a888212", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-705593104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb85e88fd4f54e0db2ff131f81137f64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba127004-dc", "ovs_interfaceid": "ba127004-dc19-4fc4-b84b-97584ac68f34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.966194] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5378c35-5571-434f-9d29-e3b6a2b5c826 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.063913] env[61906]: DEBUG nova.scheduler.client.report [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.437263] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.470974] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "refresh_cache-943c57f9-74c6-4465-9773-3fb01e78127b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.471562] env[61906]: DEBUG nova.objects.instance [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lazy-loading 'migration_context' on Instance uuid 943c57f9-74c6-4465-9773-3fb01e78127b {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.570176] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.845s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.573067] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.175s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.574641] env[61906]: INFO nova.compute.claims [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 878.590713] env[61906]: INFO nova.scheduler.client.report [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Deleted allocations for instance 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83 [ 878.975268] env[61906]: DEBUG nova.objects.base [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Object Instance<943c57f9-74c6-4465-9773-3fb01e78127b> lazy-loaded attributes: info_cache,migration_context {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 878.976491] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1906bdc-2251-4e32-baa5-e4d4586fef53 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.996945] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c9aa295-83da-43e9-b791-8101b3a34f94 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.002596] env[61906]: DEBUG oslo_vmware.api [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 879.002596] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522df1ca-7f8f-c33b-bff7-0f178dcd8555" [ 879.002596] env[61906]: _type = "Task" [ 879.002596] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.010704] env[61906]: DEBUG oslo_vmware.api [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522df1ca-7f8f-c33b-bff7-0f178dcd8555, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.101017] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7eb5b8fa-498a-41d2-a54e-a4fe3985f6ae tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "51fa46a9-b1b3-4115-b1d7-cc4f672f2e83" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.704s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.513434] env[61906]: DEBUG oslo_vmware.api [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522df1ca-7f8f-c33b-bff7-0f178dcd8555, 'name': SearchDatastore_Task, 'duration_secs': 0.027892} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.513788] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.739733] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d8f867-b486-416c-b625-23daefd07e07 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.747723] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642c322f-388d-447d-9a1f-699d20d21396 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.779252] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c916931-bed0-4b96-9125-bd3033c186df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.786852] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee9a306d-20eb-4029-a5e0-479d220dd1cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.800826] env[61906]: DEBUG nova.compute.provider_tree [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.304637] env[61906]: DEBUG nova.scheduler.client.report [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.810522] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.238s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.811088] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 880.813775] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.380s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.815619] env[61906]: INFO nova.compute.claims [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.322962] env[61906]: DEBUG nova.compute.utils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 881.325652] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 881.325934] env[61906]: DEBUG nova.network.neutron [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 881.381033] env[61906]: DEBUG nova.policy [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747d463bd48b48f0a928417917774ebd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a8ab66571abf4eb7a6431e4aa9d9dd3e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 881.651224] env[61906]: DEBUG nova.network.neutron [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Successfully created port: 6664c3db-45ba-4a24-a4a1-2b00df77b08f {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 881.825972] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 881.999026] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7f6cac-93eb-4f85-8794-2130256c932a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.006569] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7a1538-80c0-43ab-af89-d0663a9f0e20 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.037516] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-826c3aec-0820-4ea4-abea-253c147e608c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.044942] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ded43a2-609c-418e-bbdd-b5d217fe3d7f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.058262] env[61906]: DEBUG nova.compute.provider_tree [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.561883] env[61906]: DEBUG nova.scheduler.client.report [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.839170] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 883.066886] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.067485] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 883.070340] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.633s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.071999] env[61906]: INFO nova.compute.claims [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 883.169049] env[61906]: DEBUG nova.network.neutron [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Successfully updated port: 6664c3db-45ba-4a24-a4a1-2b00df77b08f {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 883.577088] env[61906]: DEBUG nova.compute.utils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.580951] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 883.581149] env[61906]: DEBUG nova.network.neutron [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 883.669774] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 883.670653] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 883.670914] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 883.671151] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 883.671316] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 883.671470] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 883.671691] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 883.671859] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 883.672246] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 883.672246] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 883.672403] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 883.675487] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "refresh_cache-d0792521-4637-4ce7-a579-71caa80ba38c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.675487] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired lock "refresh_cache-d0792521-4637-4ce7-a579-71caa80ba38c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.675487] env[61906]: DEBUG nova.network.neutron [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.677359] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb93643a-d8a6-4f11-a7df-b3bd3f99f2ed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.688518] env[61906]: DEBUG oslo_vmware.rw_handles [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529482d4-9405-b814-0797-c1922ece743b/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 883.690060] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb497d4-7dcd-4111-b0e2-f7805616a6db {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.695036] env[61906]: DEBUG nova.policy [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747d463bd48b48f0a928417917774ebd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a8ab66571abf4eb7a6431e4aa9d9dd3e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.707478] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d830abe-41ba-4192-b944-f084b1670364 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.710548] env[61906]: DEBUG oslo_vmware.rw_handles [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529482d4-9405-b814-0797-c1922ece743b/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 883.710723] env[61906]: ERROR oslo_vmware.rw_handles [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529482d4-9405-b814-0797-c1922ece743b/disk-0.vmdk due to incomplete transfer. [ 883.711737] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-02e9e06e-cc74-46ab-9201-e36265e833fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.736830] env[61906]: DEBUG oslo_vmware.rw_handles [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529482d4-9405-b814-0797-c1922ece743b/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 883.737088] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Uploaded image 629c5099-4091-4b95-8c5a-a3a2b13b491d to the Glance image server {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 883.743164] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 883.743471] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-49af35f2-e312-4063-afb4-e7447ee33582 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.750542] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 883.750542] env[61906]: value = "task-1356791" [ 883.750542] env[61906]: _type = "Task" [ 883.750542] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.759901] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356791, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.901796] env[61906]: DEBUG nova.compute.manager [req-a8a365a3-f6f0-4d8e-bccd-ad1bfba76ba5 req-9f3fc2b6-291c-4e4d-997c-dabeeb2646cd service nova] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Received event network-vif-plugged-6664c3db-45ba-4a24-a4a1-2b00df77b08f {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.901796] env[61906]: DEBUG oslo_concurrency.lockutils [req-a8a365a3-f6f0-4d8e-bccd-ad1bfba76ba5 req-9f3fc2b6-291c-4e4d-997c-dabeeb2646cd service nova] Acquiring lock "d0792521-4637-4ce7-a579-71caa80ba38c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.901796] env[61906]: DEBUG oslo_concurrency.lockutils [req-a8a365a3-f6f0-4d8e-bccd-ad1bfba76ba5 req-9f3fc2b6-291c-4e4d-997c-dabeeb2646cd service nova] Lock "d0792521-4637-4ce7-a579-71caa80ba38c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.901796] env[61906]: DEBUG oslo_concurrency.lockutils [req-a8a365a3-f6f0-4d8e-bccd-ad1bfba76ba5 req-9f3fc2b6-291c-4e4d-997c-dabeeb2646cd service nova] Lock "d0792521-4637-4ce7-a579-71caa80ba38c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.901796] env[61906]: DEBUG nova.compute.manager [req-a8a365a3-f6f0-4d8e-bccd-ad1bfba76ba5 req-9f3fc2b6-291c-4e4d-997c-dabeeb2646cd service nova] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] No waiting events found dispatching network-vif-plugged-6664c3db-45ba-4a24-a4a1-2b00df77b08f {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 883.901796] env[61906]: WARNING nova.compute.manager [req-a8a365a3-f6f0-4d8e-bccd-ad1bfba76ba5 req-9f3fc2b6-291c-4e4d-997c-dabeeb2646cd service nova] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Received unexpected event network-vif-plugged-6664c3db-45ba-4a24-a4a1-2b00df77b08f for instance with vm_state building and task_state spawning. [ 884.082282] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 884.086791] env[61906]: DEBUG nova.network.neutron [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Successfully created port: cc178483-0a7d-4037-8bd3-5d0a6a4ffefb {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 884.188543] env[61906]: DEBUG oslo_concurrency.lockutils [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.189447] env[61906]: DEBUG oslo_concurrency.lockutils [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.190560] env[61906]: INFO nova.compute.manager [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Shelving [ 884.237364] env[61906]: DEBUG nova.network.neutron [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.267549] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356791, 'name': Destroy_Task, 'duration_secs': 0.413194} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.271184] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Destroyed the VM [ 884.271621] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 884.272454] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f85aa6b9-01d6-4518-ba7e-3097b9e2ba7c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.280815] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 884.280815] env[61906]: value = "task-1356792" [ 884.280815] env[61906]: _type = "Task" [ 884.280815] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.295532] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356792, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.299673] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba2ad92-af96-43ac-b754-1884bfd8c9f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.307337] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54082d8-11e2-4f26-ac95-c631ed933642 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.340374] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b622231e-514b-49aa-a4e4-1ec52573d6fd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.350109] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb57c57e-08b2-4ebf-9350-9786ea3ae31d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.365292] env[61906]: DEBUG nova.compute.provider_tree [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.483580] env[61906]: DEBUG nova.network.neutron [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Updating instance_info_cache with network_info: [{"id": "6664c3db-45ba-4a24-a4a1-2b00df77b08f", "address": "fa:16:3e:32:e1:1f", "network": {"id": "ef77b791-3dbc-4771-b384-b07b20ffd012", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1332279450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8ab66571abf4eb7a6431e4aa9d9dd3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6664c3db-45", "ovs_interfaceid": "6664c3db-45ba-4a24-a4a1-2b00df77b08f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.705149] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 884.705149] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89717642-73b0-46fc-b0d6-d795bc650886 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.712140] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 884.712140] env[61906]: value = "task-1356793" [ 884.712140] env[61906]: _type = "Task" [ 884.712140] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.721838] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356793, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.791359] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356792, 'name': RemoveSnapshot_Task, 'duration_secs': 0.430192} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.791661] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 884.791950] env[61906]: DEBUG nova.compute.manager [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.792745] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11abfea8-b8ff-400f-a78e-6e947c55cd94 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.868716] env[61906]: DEBUG nova.scheduler.client.report [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.986759] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Releasing lock "refresh_cache-d0792521-4637-4ce7-a579-71caa80ba38c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.987178] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Instance network_info: |[{"id": "6664c3db-45ba-4a24-a4a1-2b00df77b08f", "address": "fa:16:3e:32:e1:1f", "network": {"id": "ef77b791-3dbc-4771-b384-b07b20ffd012", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1332279450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8ab66571abf4eb7a6431e4aa9d9dd3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6664c3db-45", "ovs_interfaceid": "6664c3db-45ba-4a24-a4a1-2b00df77b08f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 884.987695] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:e1:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '098df9b7-d759-47f7-b756-334848cb423b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6664c3db-45ba-4a24-a4a1-2b00df77b08f', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.995833] env[61906]: DEBUG oslo.service.loopingcall [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.996169] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.996689] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-139e882b-6d5c-4b7a-9bc5-f85bd525c5c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.018516] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 885.018516] env[61906]: value = "task-1356794" [ 885.018516] env[61906]: _type = "Task" [ 885.018516] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.034879] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356794, 'name': CreateVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.092540] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 885.121158] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 885.121450] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 885.121619] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.121830] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 885.121991] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.122538] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 885.122869] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 885.123039] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 885.123190] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 885.123442] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 885.123644] env[61906]: DEBUG nova.virt.hardware [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.124605] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0da7b81-b28f-4958-97b6-6fb56c31d86e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.132907] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0a4e63-1ee1-4d82-9c48-3f25d98b5cf1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.223026] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356793, 'name': PowerOffVM_Task, 'duration_secs': 0.289796} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.223026] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 885.224507] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea13420-72be-45bf-9303-aff3436c638d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.245916] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca6e191-271f-4c72-b12a-65ee52e74c12 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.306670] env[61906]: INFO nova.compute.manager [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Shelve offloading [ 885.308759] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 885.309110] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4f3ecff-e1a2-4d42-a838-48ee06bf7e16 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.318740] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 885.318740] env[61906]: value = "task-1356795" [ 885.318740] env[61906]: _type = "Task" [ 885.318740] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.327108] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356795, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.374855] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.375387] env[61906]: DEBUG nova.compute.manager [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 885.378449] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 5.865s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.529078] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356794, 'name': CreateVM_Task, 'duration_secs': 0.391061} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.529287] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.530112] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.530314] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.530667] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.530944] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7001947c-d4e2-4748-b44c-7fd90b1f8803 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.535401] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 885.535401] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52294e55-9287-0db7-0ac8-7b5ddefe36c1" [ 885.535401] env[61906]: _type = "Task" [ 885.535401] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.543382] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52294e55-9287-0db7-0ac8-7b5ddefe36c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.757473] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 885.757860] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5972dd17-f361-4727-bf02-603e63d5e674 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.765342] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 885.765342] env[61906]: value = "task-1356796" [ 885.765342] env[61906]: _type = "Task" [ 885.765342] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.774534] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356796, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.833811] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] VM already powered off {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 885.833811] env[61906]: DEBUG nova.compute.manager [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.835323] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0993b2c3-9483-48b0-a02b-05c2cee5ab4d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.841973] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "refresh_cache-33e288d5-9065-4606-b17a-c4c90bcbd533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.843080] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "refresh_cache-33e288d5-9065-4606-b17a-c4c90bcbd533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.843080] env[61906]: DEBUG nova.network.neutron [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 885.885835] env[61906]: DEBUG nova.compute.utils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.890801] env[61906]: DEBUG nova.compute.manager [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 885.890801] env[61906]: DEBUG nova.network.neutron [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 885.942903] env[61906]: DEBUG nova.compute.manager [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Received event network-changed-6664c3db-45ba-4a24-a4a1-2b00df77b08f {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.943144] env[61906]: DEBUG nova.compute.manager [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Refreshing instance network info cache due to event network-changed-6664c3db-45ba-4a24-a4a1-2b00df77b08f. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 885.943378] env[61906]: DEBUG oslo_concurrency.lockutils [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] Acquiring lock "refresh_cache-d0792521-4637-4ce7-a579-71caa80ba38c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.943524] env[61906]: DEBUG oslo_concurrency.lockutils [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] Acquired lock "refresh_cache-d0792521-4637-4ce7-a579-71caa80ba38c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.943730] env[61906]: DEBUG nova.network.neutron [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Refreshing network info cache for port 6664c3db-45ba-4a24-a4a1-2b00df77b08f {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 885.962423] env[61906]: DEBUG nova.policy [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73219f5e96f64d1da86a19da9b18b06a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5533be74874b4094b7ef0f98a121f60b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 885.993824] env[61906]: DEBUG nova.network.neutron [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Successfully updated port: cc178483-0a7d-4037-8bd3-5d0a6a4ffefb {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.050175] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52294e55-9287-0db7-0ac8-7b5ddefe36c1, 'name': SearchDatastore_Task, 'duration_secs': 0.010491} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.050175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.050175] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 886.051073] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.051073] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.051073] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.051073] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-913e8d4c-40e1-429e-8fd0-8c46e1ebbb14 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.060105] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.060105] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 886.063310] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-670289a8-f65d-4112-9c88-8ba11a1e0891 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.070222] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 886.070222] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529920a8-e03e-5d6d-8870-d5328a832fbd" [ 886.070222] env[61906]: _type = "Task" [ 886.070222] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.078962] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529920a8-e03e-5d6d-8870-d5328a832fbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.107352] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a31f3793-e6be-4127-ad8f-71d970885622 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.114665] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d817fb-3cc5-4daf-959a-39bcf1564667 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.150029] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d93ef3a-818b-413d-afb7-3287630d4f49 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.155388] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-178df553-ae65-4ec8-87d2-9c00cedb6cc1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.171132] env[61906]: DEBUG nova.compute.provider_tree [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.275698] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356796, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.335244] env[61906]: DEBUG oslo_vmware.rw_handles [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a63af9-3a7a-0cce-df05-b66591c7af0f/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 886.336991] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3290f25-fe9b-4cfb-8ccf-f181c71417d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.343059] env[61906]: DEBUG oslo_vmware.rw_handles [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a63af9-3a7a-0cce-df05-b66591c7af0f/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 886.343246] env[61906]: ERROR oslo_vmware.rw_handles [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a63af9-3a7a-0cce-df05-b66591c7af0f/disk-0.vmdk due to incomplete transfer. [ 886.343480] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a757d043-4489-4a16-9c2e-1c1e0945a1c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.345706] env[61906]: DEBUG nova.network.neutron [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Successfully created port: 9a08ec03-408f-4d57-842d-3f68f53dcb29 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.361333] env[61906]: DEBUG oslo_vmware.rw_handles [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a63af9-3a7a-0cce-df05-b66591c7af0f/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 886.361553] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Uploaded image cc54d929-47af-46bb-8c53-a38fb410a7cc to the Glance image server {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 886.363407] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 886.363678] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-60b3e0fe-6807-4a66-aa30-6ffee4fc4649 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.370967] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 886.370967] env[61906]: value = "task-1356797" [ 886.370967] env[61906]: _type = "Task" [ 886.370967] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.378667] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356797, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.394842] env[61906]: DEBUG nova.compute.manager [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 886.498822] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "refresh_cache-03bd64e4-16fc-4659-a428-a2d9e7205b81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.498999] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired lock "refresh_cache-03bd64e4-16fc-4659-a428-a2d9e7205b81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.499271] env[61906]: DEBUG nova.network.neutron [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.586244] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529920a8-e03e-5d6d-8870-d5328a832fbd, 'name': SearchDatastore_Task, 'duration_secs': 0.009444} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.587122] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e698c427-b613-4cd9-90a1-9de531eff9dc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.592533] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 886.592533] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521c7d84-2841-f0e0-9d1f-c6dd1a65424e" [ 886.592533] env[61906]: _type = "Task" [ 886.592533] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.600980] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521c7d84-2841-f0e0-9d1f-c6dd1a65424e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.608681] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.608934] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.675253] env[61906]: DEBUG nova.scheduler.client.report [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.705647] env[61906]: DEBUG nova.network.neutron [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Updated VIF entry in instance network info cache for port 6664c3db-45ba-4a24-a4a1-2b00df77b08f. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 886.706010] env[61906]: DEBUG nova.network.neutron [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Updating instance_info_cache with network_info: [{"id": "6664c3db-45ba-4a24-a4a1-2b00df77b08f", "address": "fa:16:3e:32:e1:1f", "network": {"id": "ef77b791-3dbc-4771-b384-b07b20ffd012", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1332279450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8ab66571abf4eb7a6431e4aa9d9dd3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6664c3db-45", "ovs_interfaceid": "6664c3db-45ba-4a24-a4a1-2b00df77b08f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.756108] env[61906]: DEBUG nova.network.neutron [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Updating instance_info_cache with network_info: [{"id": "a88ac5a5-d8da-4ebb-a745-ecc89275cb78", "address": "fa:16:3e:6e:19:ce", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa88ac5a5-d8", "ovs_interfaceid": "a88ac5a5-d8da-4ebb-a745-ecc89275cb78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.775745] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356796, 'name': CreateSnapshot_Task, 'duration_secs': 0.821523} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.776061] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 886.776999] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3725f962-4b0c-4ea4-b3d8-ae0322c9e3a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.885978] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356797, 'name': Destroy_Task} progress is 33%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.029648] env[61906]: DEBUG nova.network.neutron [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.105417] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521c7d84-2841-f0e0-9d1f-c6dd1a65424e, 'name': SearchDatastore_Task, 'duration_secs': 0.009754} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.105692] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.105962] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] d0792521-4637-4ce7-a579-71caa80ba38c/d0792521-4637-4ce7-a579-71caa80ba38c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 887.106241] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f119a4a7-3ea5-4593-b138-9ce3eaccd455 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.111956] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 887.111956] env[61906]: value = "task-1356798" [ 887.111956] env[61906]: _type = "Task" [ 887.111956] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.116659] env[61906]: DEBUG nova.compute.manager [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 887.123523] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356798, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.175187] env[61906]: DEBUG nova.network.neutron [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Updating instance_info_cache with network_info: [{"id": "cc178483-0a7d-4037-8bd3-5d0a6a4ffefb", "address": "fa:16:3e:2f:71:a5", "network": {"id": "ef77b791-3dbc-4771-b384-b07b20ffd012", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1332279450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8ab66571abf4eb7a6431e4aa9d9dd3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc178483-0a", "ovs_interfaceid": "cc178483-0a7d-4037-8bd3-5d0a6a4ffefb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.208870] env[61906]: DEBUG oslo_concurrency.lockutils [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] Releasing lock "refresh_cache-d0792521-4637-4ce7-a579-71caa80ba38c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.209126] env[61906]: DEBUG nova.compute.manager [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Received event network-vif-plugged-cc178483-0a7d-4037-8bd3-5d0a6a4ffefb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.209343] env[61906]: DEBUG oslo_concurrency.lockutils [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] Acquiring lock "03bd64e4-16fc-4659-a428-a2d9e7205b81-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.209557] env[61906]: DEBUG oslo_concurrency.lockutils [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] Lock "03bd64e4-16fc-4659-a428-a2d9e7205b81-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.209783] env[61906]: DEBUG oslo_concurrency.lockutils [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] Lock "03bd64e4-16fc-4659-a428-a2d9e7205b81-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.209892] env[61906]: DEBUG nova.compute.manager [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] No waiting events found dispatching network-vif-plugged-cc178483-0a7d-4037-8bd3-5d0a6a4ffefb {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 887.210074] env[61906]: WARNING nova.compute.manager [req-80bac21b-e3fb-4b24-9945-97806a250c61 req-67611656-ce55-4622-99da-cfb83fb19151 service nova] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Received unexpected event network-vif-plugged-cc178483-0a7d-4037-8bd3-5d0a6a4ffefb for instance with vm_state building and task_state spawning. [ 887.259027] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "refresh_cache-33e288d5-9065-4606-b17a-c4c90bcbd533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.294766] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 887.295180] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ffe778e1-7605-403e-bb48-303b9a1244ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.306553] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 887.306553] env[61906]: value = "task-1356799" [ 887.306553] env[61906]: _type = "Task" [ 887.306553] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.317431] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356799, 'name': CloneVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.382018] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356797, 'name': Destroy_Task, 'duration_secs': 0.677455} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.382402] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Destroyed the VM [ 887.382653] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 887.382929] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-71bce162-5c52-4925-b784-0916c1bd8223 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.389179] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 887.389179] env[61906]: value = "task-1356800" [ 887.389179] env[61906]: _type = "Task" [ 887.389179] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.397934] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356800, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.407654] env[61906]: DEBUG nova.compute.manager [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 887.447433] env[61906]: DEBUG nova.virt.hardware [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.447433] env[61906]: DEBUG nova.virt.hardware [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.447433] env[61906]: DEBUG nova.virt.hardware [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.447433] env[61906]: DEBUG nova.virt.hardware [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.447433] env[61906]: DEBUG nova.virt.hardware [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.447433] env[61906]: DEBUG nova.virt.hardware [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.447999] env[61906]: DEBUG nova.virt.hardware [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.448930] env[61906]: DEBUG nova.virt.hardware [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.448930] env[61906]: DEBUG nova.virt.hardware [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.448930] env[61906]: DEBUG nova.virt.hardware [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.449147] env[61906]: DEBUG nova.virt.hardware [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.450107] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e796869-d20c-4499-b960-ca18d11b584d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.460249] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c49c4749-364a-447e-a331-46a896b9c30a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.628445] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356798, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.649516] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.665388] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "e35580b3-7177-46fd-bb86-b013efbf4911" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.665643] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "e35580b3-7177-46fd-bb86-b013efbf4911" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.676902] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 887.677616] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Releasing lock "refresh_cache-03bd64e4-16fc-4659-a428-a2d9e7205b81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.678009] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Instance network_info: |[{"id": "cc178483-0a7d-4037-8bd3-5d0a6a4ffefb", "address": "fa:16:3e:2f:71:a5", "network": {"id": "ef77b791-3dbc-4771-b384-b07b20ffd012", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1332279450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8ab66571abf4eb7a6431e4aa9d9dd3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc178483-0a", "ovs_interfaceid": "cc178483-0a7d-4037-8bd3-5d0a6a4ffefb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 887.678900] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f19b84-d29d-4ff6-8020-7ad643b4a280 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.681975] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:71:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '098df9b7-d759-47f7-b756-334848cb423b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc178483-0a7d-4037-8bd3-5d0a6a4ffefb', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.689587] env[61906]: DEBUG oslo.service.loopingcall [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.691051] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.313s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.694518] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 887.695506] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.046s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.697078] env[61906]: INFO nova.compute.claims [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.700028] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bab84906-f0eb-4b6b-87c4-654c71bea940 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.717514] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 887.718238] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-62056f19-5548-4eb6-b9a7-06116c8a4896 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.724065] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.724065] env[61906]: value = "task-1356801" [ 887.724065] env[61906]: _type = "Task" [ 887.724065] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.732188] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356801, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.798394] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 887.798799] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 887.798939] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleting the datastore file [datastore2] 33e288d5-9065-4606-b17a-c4c90bcbd533 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.799228] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df4dfd29-627a-4478-847d-9def807b6dfd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.805649] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 887.805649] env[61906]: value = "task-1356803" [ 887.805649] env[61906]: _type = "Task" [ 887.805649] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.818666] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356803, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.821872] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356799, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.899069] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356800, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.983971] env[61906]: DEBUG nova.compute.manager [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Received event network-changed-cc178483-0a7d-4037-8bd3-5d0a6a4ffefb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.984178] env[61906]: DEBUG nova.compute.manager [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Refreshing instance network info cache due to event network-changed-cc178483-0a7d-4037-8bd3-5d0a6a4ffefb. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 887.984405] env[61906]: DEBUG oslo_concurrency.lockutils [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] Acquiring lock "refresh_cache-03bd64e4-16fc-4659-a428-a2d9e7205b81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.984604] env[61906]: DEBUG oslo_concurrency.lockutils [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] Acquired lock "refresh_cache-03bd64e4-16fc-4659-a428-a2d9e7205b81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.984836] env[61906]: DEBUG nova.network.neutron [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Refreshing network info cache for port cc178483-0a7d-4037-8bd3-5d0a6a4ffefb {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 888.071403] env[61906]: DEBUG nova.network.neutron [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Successfully updated port: 9a08ec03-408f-4d57-842d-3f68f53dcb29 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.124118] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356798, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.168456] env[61906]: DEBUG nova.compute.manager [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 888.236940] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356801, 'name': CreateVM_Task, 'duration_secs': 0.411899} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.237136] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 888.237908] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.238121] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.238596] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.238890] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-333ea2e0-06dd-4182-850b-72c72a7e262a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.243989] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 888.243989] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529ee24b-831a-f64a-6be7-ef128c53bc60" [ 888.243989] env[61906]: _type = "Task" [ 888.243989] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.255088] env[61906]: INFO nova.scheduler.client.report [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Deleted allocation for migration 0fb2c22f-a902-4138-8a6b-071a4253a8d9 [ 888.255964] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529ee24b-831a-f64a-6be7-ef128c53bc60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.319854] env[61906]: DEBUG oslo_vmware.api [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356803, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190289} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.324379] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.324379] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 888.324624] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 888.326714] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356799, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.346702] env[61906]: INFO nova.scheduler.client.report [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleted allocations for instance 33e288d5-9065-4606-b17a-c4c90bcbd533 [ 888.399252] env[61906]: DEBUG oslo_vmware.api [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356800, 'name': RemoveSnapshot_Task, 'duration_secs': 0.851396} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.400066] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 888.400066] env[61906]: INFO nova.compute.manager [None req-78343d89-e04e-4b3e-9006-4b0784ad5c7f tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Took 15.88 seconds to snapshot the instance on the hypervisor. [ 888.573729] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.573879] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.574044] env[61906]: DEBUG nova.network.neutron [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.623729] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356798, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.695447] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.696455] env[61906]: DEBUG nova.network.neutron [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Updated VIF entry in instance network info cache for port cc178483-0a7d-4037-8bd3-5d0a6a4ffefb. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 888.696898] env[61906]: DEBUG nova.network.neutron [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Updating instance_info_cache with network_info: [{"id": "cc178483-0a7d-4037-8bd3-5d0a6a4ffefb", "address": "fa:16:3e:2f:71:a5", "network": {"id": "ef77b791-3dbc-4771-b384-b07b20ffd012", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1332279450-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a8ab66571abf4eb7a6431e4aa9d9dd3e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "098df9b7-d759-47f7-b756-334848cb423b", "external-id": "nsx-vlan-transportzone-765", "segmentation_id": 765, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc178483-0a", "ovs_interfaceid": "cc178483-0a7d-4037-8bd3-5d0a6a4ffefb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.753521] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529ee24b-831a-f64a-6be7-ef128c53bc60, 'name': SearchDatastore_Task, 'duration_secs': 0.008624} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.756048] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.756306] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.756551] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.756739] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.756938] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.757507] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c4faab5-f878-427d-adb2-0e0cbf982421 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.759813] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c625f75-fa8a-40fa-99b2-92d70392560d tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "943c57f9-74c6-4465-9773-3fb01e78127b" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 12.661s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.766882] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.767087] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 888.770115] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a76d145-f85a-4f2c-8d91-6be2a6e866c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.776456] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 888.776456] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5248b0a3-250d-744a-fbec-5d0b409ef685" [ 888.776456] env[61906]: _type = "Task" [ 888.776456] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.786613] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5248b0a3-250d-744a-fbec-5d0b409ef685, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.822281] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356799, 'name': CloneVM_Task, 'duration_secs': 1.415493} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.824573] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Created linked-clone VM from snapshot [ 888.825490] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7064fe8-fc37-4da1-a09f-b5e69e5d577f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.833177] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Uploading image 4e467722-4f47-4b26-9f34-13132817ea0e {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 888.850710] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.857958] env[61906]: DEBUG oslo_vmware.rw_handles [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 888.857958] env[61906]: value = "vm-289005" [ 888.857958] env[61906]: _type = "VirtualMachine" [ 888.857958] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 888.858536] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c3cbd183-d7f3-4431-a6cc-37a74783f167 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.865755] env[61906]: DEBUG oslo_vmware.rw_handles [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lease: (returnval){ [ 888.865755] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a6b404-998d-570f-6fbc-0d0b5a52afda" [ 888.865755] env[61906]: _type = "HttpNfcLease" [ 888.865755] env[61906]: } obtained for exporting VM: (result){ [ 888.865755] env[61906]: value = "vm-289005" [ 888.865755] env[61906]: _type = "VirtualMachine" [ 888.865755] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 888.865755] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the lease: (returnval){ [ 888.865755] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a6b404-998d-570f-6fbc-0d0b5a52afda" [ 888.865755] env[61906]: _type = "HttpNfcLease" [ 888.865755] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 888.874387] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 888.874387] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a6b404-998d-570f-6fbc-0d0b5a52afda" [ 888.874387] env[61906]: _type = "HttpNfcLease" [ 888.874387] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 888.880399] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ea997e-a1f6-4c26-a802-816d03b933a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.887261] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fbd070-de8f-4ec9-be5d-6ea863817575 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.921012] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4d0583-1d10-48b6-81e7-e4f6582ef57a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.929358] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c8025f-a9d3-4d26-89ea-5a06c9baa3e2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.948282] env[61906]: DEBUG nova.compute.provider_tree [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.112518] env[61906]: DEBUG nova.network.neutron [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.123772] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356798, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.515192} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.124064] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] d0792521-4637-4ce7-a579-71caa80ba38c/d0792521-4637-4ce7-a579-71caa80ba38c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 889.124286] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.124544] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1c18ea7-153f-4016-a6ba-facb25293ee0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.130998] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 889.130998] env[61906]: value = "task-1356805" [ 889.130998] env[61906]: _type = "Task" [ 889.130998] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.143299] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356805, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.202643] env[61906]: DEBUG oslo_concurrency.lockutils [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] Releasing lock "refresh_cache-03bd64e4-16fc-4659-a428-a2d9e7205b81" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.202935] env[61906]: DEBUG nova.compute.manager [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Received event network-vif-unplugged-a88ac5a5-d8da-4ebb-a745-ecc89275cb78 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.203161] env[61906]: DEBUG oslo_concurrency.lockutils [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] Acquiring lock "33e288d5-9065-4606-b17a-c4c90bcbd533-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.203375] env[61906]: DEBUG oslo_concurrency.lockutils [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] Lock "33e288d5-9065-4606-b17a-c4c90bcbd533-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.203545] env[61906]: DEBUG oslo_concurrency.lockutils [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] Lock "33e288d5-9065-4606-b17a-c4c90bcbd533-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.203716] env[61906]: DEBUG nova.compute.manager [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] No waiting events found dispatching network-vif-unplugged-a88ac5a5-d8da-4ebb-a745-ecc89275cb78 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 889.203893] env[61906]: WARNING nova.compute.manager [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Received unexpected event network-vif-unplugged-a88ac5a5-d8da-4ebb-a745-ecc89275cb78 for instance with vm_state shelved and task_state shelving_offloading. [ 889.204076] env[61906]: DEBUG nova.compute.manager [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Received event network-changed-a88ac5a5-d8da-4ebb-a745-ecc89275cb78 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.204238] env[61906]: DEBUG nova.compute.manager [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Refreshing instance network info cache due to event network-changed-a88ac5a5-d8da-4ebb-a745-ecc89275cb78. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 889.204425] env[61906]: DEBUG oslo_concurrency.lockutils [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] Acquiring lock "refresh_cache-33e288d5-9065-4606-b17a-c4c90bcbd533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.204658] env[61906]: DEBUG oslo_concurrency.lockutils [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] Acquired lock "refresh_cache-33e288d5-9065-4606-b17a-c4c90bcbd533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.204847] env[61906]: DEBUG nova.network.neutron [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Refreshing network info cache for port a88ac5a5-d8da-4ebb-a745-ecc89275cb78 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 889.286940] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5248b0a3-250d-744a-fbec-5d0b409ef685, 'name': SearchDatastore_Task, 'duration_secs': 0.008471} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.287789] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd6140e4-1a06-4eb2-b0c0-cb1e6a540ec0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.293529] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 889.293529] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5293eb9d-12c5-869b-0a56-d131e3baa726" [ 889.293529] env[61906]: _type = "Task" [ 889.293529] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.301286] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5293eb9d-12c5-869b-0a56-d131e3baa726, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.320641] env[61906]: DEBUG nova.network.neutron [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updating instance_info_cache with network_info: [{"id": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "address": "fa:16:3e:b8:da:a6", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a08ec03-40", "ovs_interfaceid": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.374554] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 889.374554] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a6b404-998d-570f-6fbc-0d0b5a52afda" [ 889.374554] env[61906]: _type = "HttpNfcLease" [ 889.374554] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 889.374882] env[61906]: DEBUG oslo_vmware.rw_handles [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 889.374882] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a6b404-998d-570f-6fbc-0d0b5a52afda" [ 889.374882] env[61906]: _type = "HttpNfcLease" [ 889.374882] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 889.375610] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73026136-7deb-4820-b19f-1e6fd37e794e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.382898] env[61906]: DEBUG oslo_vmware.rw_handles [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526c36c8-32c4-857d-1fd0-2b827cb61bf7/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 889.383096] env[61906]: DEBUG oslo_vmware.rw_handles [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526c36c8-32c4-857d-1fd0-2b827cb61bf7/disk-0.vmdk for reading. {{(pid=61906) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 889.451401] env[61906]: DEBUG nova.scheduler.client.report [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 889.470607] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f1e5b0ce-cfbb-4c10-9c1d-2861c63a7dac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.641907] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356805, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056944} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.642241] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 889.643035] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46d94c19-4195-4cea-9616-ca9ba9d84bc1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.665591] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] d0792521-4637-4ce7-a579-71caa80ba38c/d0792521-4637-4ce7-a579-71caa80ba38c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.665591] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f4e64d0-c2d9-4b37-9208-20fce170efca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.687728] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 889.687728] env[61906]: value = "task-1356806" [ 889.687728] env[61906]: _type = "Task" [ 889.687728] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.695944] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356806, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.807535] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5293eb9d-12c5-869b-0a56-d131e3baa726, 'name': SearchDatastore_Task, 'duration_secs': 0.010183} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.807995] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.808353] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 03bd64e4-16fc-4659-a428-a2d9e7205b81/03bd64e4-16fc-4659-a428-a2d9e7205b81.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 889.808749] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-87f85142-1bb0-4ffa-b916-ebc9dae1cf96 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.815884] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 889.815884] env[61906]: value = "task-1356807" [ 889.815884] env[61906]: _type = "Task" [ 889.815884] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.824344] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.824683] env[61906]: DEBUG nova.compute.manager [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Instance network_info: |[{"id": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "address": "fa:16:3e:b8:da:a6", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a08ec03-40", "ovs_interfaceid": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 889.824976] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356807, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.827556] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:da:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a08ec03-408f-4d57-842d-3f68f53dcb29', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.835610] env[61906]: DEBUG oslo.service.loopingcall [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.836691] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 889.837460] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83521c6a-3772-4f6f-bde7-690af311709b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.858262] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.858262] env[61906]: value = "task-1356808" [ 889.858262] env[61906]: _type = "Task" [ 889.858262] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.867310] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356808, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.956431] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.957129] env[61906]: DEBUG nova.compute.manager [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 889.961815] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.265s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.962484] env[61906]: INFO nova.compute.claims [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.995173] env[61906]: DEBUG nova.network.neutron [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Updated VIF entry in instance network info cache for port a88ac5a5-d8da-4ebb-a745-ecc89275cb78. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 889.995583] env[61906]: DEBUG nova.network.neutron [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Updating instance_info_cache with network_info: [{"id": "a88ac5a5-d8da-4ebb-a745-ecc89275cb78", "address": "fa:16:3e:6e:19:ce", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": null, "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapa88ac5a5-d8", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.010959] env[61906]: DEBUG nova.compute.manager [req-6b87a000-ca8d-45a8-82bb-8efef031ae29 req-d01d9c20-6810-4caf-a301-f8beca479975 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Received event network-changed-9a08ec03-408f-4d57-842d-3f68f53dcb29 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.011494] env[61906]: DEBUG nova.compute.manager [req-6b87a000-ca8d-45a8-82bb-8efef031ae29 req-d01d9c20-6810-4caf-a301-f8beca479975 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Refreshing instance network info cache due to event network-changed-9a08ec03-408f-4d57-842d-3f68f53dcb29. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 890.011749] env[61906]: DEBUG oslo_concurrency.lockutils [req-6b87a000-ca8d-45a8-82bb-8efef031ae29 req-d01d9c20-6810-4caf-a301-f8beca479975 service nova] Acquiring lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.011915] env[61906]: DEBUG oslo_concurrency.lockutils [req-6b87a000-ca8d-45a8-82bb-8efef031ae29 req-d01d9c20-6810-4caf-a301-f8beca479975 service nova] Acquired lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.012541] env[61906]: DEBUG nova.network.neutron [req-6b87a000-ca8d-45a8-82bb-8efef031ae29 req-d01d9c20-6810-4caf-a301-f8beca479975 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Refreshing network info cache for port 9a08ec03-408f-4d57-842d-3f68f53dcb29 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 890.198968] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356806, 'name': ReconfigVM_Task, 'duration_secs': 0.303519} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.199451] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Reconfigured VM instance instance-00000053 to attach disk [datastore2] d0792521-4637-4ce7-a579-71caa80ba38c/d0792521-4637-4ce7-a579-71caa80ba38c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.200831] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-29029e28-9a6a-414a-8410-d46670e8e8ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.208759] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 890.208759] env[61906]: value = "task-1356809" [ 890.208759] env[61906]: _type = "Task" [ 890.208759] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.219042] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356809, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.263217] env[61906]: DEBUG oslo_concurrency.lockutils [None req-babcfaa3-0806-4d41-bc6d-297a421247d2 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "33e288d5-9065-4606-b17a-c4c90bcbd533" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.326413] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356807, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.369221] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356808, 'name': CreateVM_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.467285] env[61906]: DEBUG nova.compute.utils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 890.471293] env[61906]: DEBUG nova.compute.manager [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 890.471293] env[61906]: DEBUG nova.network.neutron [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 890.492779] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "943c57f9-74c6-4465-9773-3fb01e78127b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.493498] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "943c57f9-74c6-4465-9773-3fb01e78127b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.493729] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "943c57f9-74c6-4465-9773-3fb01e78127b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.494160] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "943c57f9-74c6-4465-9773-3fb01e78127b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.494450] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "943c57f9-74c6-4465-9773-3fb01e78127b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.496714] env[61906]: INFO nova.compute.manager [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Terminating instance [ 890.498590] env[61906]: DEBUG nova.compute.manager [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 890.498794] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 890.499565] env[61906]: DEBUG oslo_concurrency.lockutils [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] Releasing lock "refresh_cache-33e288d5-9065-4606-b17a-c4c90bcbd533" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.499916] env[61906]: DEBUG nova.compute.manager [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Received event network-vif-plugged-9a08ec03-408f-4d57-842d-3f68f53dcb29 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.500242] env[61906]: DEBUG oslo_concurrency.lockutils [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] Acquiring lock "99e5c55d-79ae-47a6-8500-79fc68291650-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.500467] env[61906]: DEBUG oslo_concurrency.lockutils [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] Lock "99e5c55d-79ae-47a6-8500-79fc68291650-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.500640] env[61906]: DEBUG oslo_concurrency.lockutils [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] Lock "99e5c55d-79ae-47a6-8500-79fc68291650-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.500825] env[61906]: DEBUG nova.compute.manager [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] No waiting events found dispatching network-vif-plugged-9a08ec03-408f-4d57-842d-3f68f53dcb29 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 890.501009] env[61906]: WARNING nova.compute.manager [req-68a01be8-dcac-4ab4-b47e-c0087d339080 req-3ef29ab7-95c1-428e-afff-f1591743ebc3 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Received unexpected event network-vif-plugged-9a08ec03-408f-4d57-842d-3f68f53dcb29 for instance with vm_state building and task_state spawning. [ 890.501926] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d46ecaa-fbc0-4240-893b-05ac2b3f2f96 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.509868] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.510228] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44d4aa15-7326-4f47-8846-c896f6e1908a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.517867] env[61906]: DEBUG oslo_vmware.api [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 890.517867] env[61906]: value = "task-1356810" [ 890.517867] env[61906]: _type = "Task" [ 890.517867] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.525344] env[61906]: DEBUG nova.policy [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9dba4398586641c9adfe209ea853538f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17447d31611a46bbabae82a3ea2a2f22', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 890.531486] env[61906]: DEBUG oslo_vmware.api [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356810, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.722489] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356809, 'name': Rename_Task, 'duration_secs': 0.268668} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.722934] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 890.723280] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46fadf79-77f0-4fc3-8485-64aa4557b953 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.737272] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 890.737272] env[61906]: value = "task-1356811" [ 890.737272] env[61906]: _type = "Task" [ 890.737272] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.745735] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356811, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.828095] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356807, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512049} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.828719] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 03bd64e4-16fc-4659-a428-a2d9e7205b81/03bd64e4-16fc-4659-a428-a2d9e7205b81.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 890.829152] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.829510] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-732f2f6c-657e-4892-990b-0b34245381e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.836081] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 890.836081] env[61906]: value = "task-1356812" [ 890.836081] env[61906]: _type = "Task" [ 890.836081] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.850667] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356812, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.867943] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356808, 'name': CreateVM_Task, 'duration_secs': 0.624738} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.868514] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 890.869384] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.869572] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.870023] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.870514] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3eddf63a-a923-48a7-8975-96c8df0e5344 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.875475] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 890.875475] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52180c1d-309b-3024-f7e5-deb263084fc8" [ 890.875475] env[61906]: _type = "Task" [ 890.875475] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.883975] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52180c1d-309b-3024-f7e5-deb263084fc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.971755] env[61906]: DEBUG nova.compute.manager [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 890.975832] env[61906]: DEBUG nova.network.neutron [req-6b87a000-ca8d-45a8-82bb-8efef031ae29 req-d01d9c20-6810-4caf-a301-f8beca479975 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updated VIF entry in instance network info cache for port 9a08ec03-408f-4d57-842d-3f68f53dcb29. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 890.976276] env[61906]: DEBUG nova.network.neutron [req-6b87a000-ca8d-45a8-82bb-8efef031ae29 req-d01d9c20-6810-4caf-a301-f8beca479975 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updating instance_info_cache with network_info: [{"id": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "address": "fa:16:3e:b8:da:a6", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a08ec03-40", "ovs_interfaceid": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.005098] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "bc773559-fe14-485f-8bdc-1d4aab186521" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.005339] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "bc773559-fe14-485f-8bdc-1d4aab186521" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.033202] env[61906]: DEBUG oslo_vmware.api [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356810, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.090095] env[61906]: DEBUG nova.network.neutron [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Successfully created port: e4fb390e-8245-4c57-a3b3-ef4556435df6 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 891.172601] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00bf017-c0a8-41e6-b17b-dd6f7fb180b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.181207] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1760e6d-a094-430f-8ee9-d1c52e8923b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.216798] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14b5b27-509c-4939-bd7c-243609cac215 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.224929] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a602747-1768-45d9-8311-4191e9f0b2a3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.240633] env[61906]: DEBUG nova.compute.provider_tree [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.252525] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356811, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.345189] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356812, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065303} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.345552] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 891.346483] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f42c900-0450-44e8-a6d1-2c28a75d0c7a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.369205] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 03bd64e4-16fc-4659-a428-a2d9e7205b81/03bd64e4-16fc-4659-a428-a2d9e7205b81.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.369695] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d95504ea-6482-4112-890e-548a69073d22 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.393606] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52180c1d-309b-3024-f7e5-deb263084fc8, 'name': SearchDatastore_Task, 'duration_secs': 0.007741} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.395303] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.395649] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.395976] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.396169] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.396454] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.397042] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 891.397042] env[61906]: value = "task-1356813" [ 891.397042] env[61906]: _type = "Task" [ 891.397042] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.397309] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1be8d381-09a1-4dde-8983-83c560e18c96 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.408887] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356813, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.410601] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.411245] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.412351] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7d95ae5-2014-497a-865a-ed3cd4a18ba4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.417584] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 891.417584] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5201a8e9-da18-80d2-3ddc-83499591e27c" [ 891.417584] env[61906]: _type = "Task" [ 891.417584] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.425743] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5201a8e9-da18-80d2-3ddc-83499591e27c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.485784] env[61906]: DEBUG oslo_concurrency.lockutils [req-6b87a000-ca8d-45a8-82bb-8efef031ae29 req-d01d9c20-6810-4caf-a301-f8beca479975 service nova] Releasing lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.508636] env[61906]: DEBUG nova.compute.manager [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 891.529780] env[61906]: DEBUG oslo_vmware.api [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356810, 'name': PowerOffVM_Task, 'duration_secs': 0.618308} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.530741] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.531178] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 891.531479] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d64f14a-4a2f-4e33-aa62-088052e07960 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.601565] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 891.601900] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 891.602220] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Deleting the datastore file [datastore1] 943c57f9-74c6-4465-9773-3fb01e78127b {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.602579] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb83b782-d04f-4009-9640-e4588a74ad9a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.610031] env[61906]: DEBUG oslo_vmware.api [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 891.610031] env[61906]: value = "task-1356815" [ 891.610031] env[61906]: _type = "Task" [ 891.610031] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.620449] env[61906]: DEBUG oslo_vmware.api [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356815, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.750362] env[61906]: DEBUG nova.scheduler.client.report [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.753676] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356811, 'name': PowerOnVM_Task, 'duration_secs': 0.571737} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.754217] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 891.755629] env[61906]: INFO nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Took 8.92 seconds to spawn the instance on the hypervisor. [ 891.755629] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 891.755791] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8230677-5e69-4e83-b4bc-edd7057b1d96 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.909156] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356813, 'name': ReconfigVM_Task, 'duration_secs': 0.307039} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.909467] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 03bd64e4-16fc-4659-a428-a2d9e7205b81/03bd64e4-16fc-4659-a428-a2d9e7205b81.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.910352] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c35a3897-43ac-43cf-ab16-0a648f3974d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.915778] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 891.915778] env[61906]: value = "task-1356816" [ 891.915778] env[61906]: _type = "Task" [ 891.915778] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.931734] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356816, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.936560] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5201a8e9-da18-80d2-3ddc-83499591e27c, 'name': SearchDatastore_Task, 'duration_secs': 0.013471} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.937813] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-090aba09-88ad-4b8f-8bd0-12329bc73225 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.944434] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 891.944434] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529c0e78-d6bb-56d9-5808-12ccc4156c84" [ 891.944434] env[61906]: _type = "Task" [ 891.944434] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.955103] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529c0e78-d6bb-56d9-5808-12ccc4156c84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.989274] env[61906]: DEBUG nova.compute.manager [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 892.017740] env[61906]: DEBUG nova.virt.hardware [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 892.018114] env[61906]: DEBUG nova.virt.hardware [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 892.018299] env[61906]: DEBUG nova.virt.hardware [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 892.018509] env[61906]: DEBUG nova.virt.hardware [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 892.019234] env[61906]: DEBUG nova.virt.hardware [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 892.019234] env[61906]: DEBUG nova.virt.hardware [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 892.019234] env[61906]: DEBUG nova.virt.hardware [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 892.019393] env[61906]: DEBUG nova.virt.hardware [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 892.019507] env[61906]: DEBUG nova.virt.hardware [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 892.019694] env[61906]: DEBUG nova.virt.hardware [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 892.019880] env[61906]: DEBUG nova.virt.hardware [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 892.020828] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4514093-1313-45aa-b245-9ba0c481719c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.031699] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d02c2f3-9918-43d1-ad7e-9faafb829ffa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.047182] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.119987] env[61906]: DEBUG oslo_vmware.api [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356815, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.248164} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.120332] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.120566] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 892.120785] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 892.120982] env[61906]: INFO nova.compute.manager [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Took 1.62 seconds to destroy the instance on the hypervisor. [ 892.121284] env[61906]: DEBUG oslo.service.loopingcall [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.121503] env[61906]: DEBUG nova.compute.manager [-] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 892.121620] env[61906]: DEBUG nova.network.neutron [-] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 892.255697] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.295s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.256551] env[61906]: DEBUG nova.compute.manager [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 892.260558] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.410s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.263534] env[61906]: DEBUG nova.objects.instance [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lazy-loading 'resources' on Instance uuid 33e288d5-9065-4606-b17a-c4c90bcbd533 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.277028] env[61906]: INFO nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Took 15.90 seconds to build instance. [ 892.358512] env[61906]: DEBUG nova.compute.manager [req-63b33a54-2c32-48f2-abe0-83f3c01d5c51 req-cba6f7d8-6e6d-469b-8be7-2a2379d032f1 service nova] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Received event network-vif-deleted-ba127004-dc19-4fc4-b84b-97584ac68f34 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.358932] env[61906]: INFO nova.compute.manager [req-63b33a54-2c32-48f2-abe0-83f3c01d5c51 req-cba6f7d8-6e6d-469b-8be7-2a2379d032f1 service nova] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Neutron deleted interface ba127004-dc19-4fc4-b84b-97584ac68f34; detaching it from the instance and deleting it from the info cache [ 892.358932] env[61906]: DEBUG nova.network.neutron [req-63b33a54-2c32-48f2-abe0-83f3c01d5c51 req-cba6f7d8-6e6d-469b-8be7-2a2379d032f1 service nova] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.429850] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356816, 'name': Rename_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.456243] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529c0e78-d6bb-56d9-5808-12ccc4156c84, 'name': SearchDatastore_Task, 'duration_secs': 0.016687} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.456730] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.457724] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 99e5c55d-79ae-47a6-8500-79fc68291650/99e5c55d-79ae-47a6-8500-79fc68291650.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.458089] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07779684-a57e-4d91-843d-2d1a6d1c5461 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.465010] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 892.465010] env[61906]: value = "task-1356817" [ 892.465010] env[61906]: _type = "Task" [ 892.465010] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.473790] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356817, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.569061] env[61906]: DEBUG nova.compute.manager [req-6c97b3f7-fd19-425b-a42b-48f788087889 req-ba2f515f-15d9-4e33-8a85-025e40be019f service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Received event network-vif-plugged-e4fb390e-8245-4c57-a3b3-ef4556435df6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.569313] env[61906]: DEBUG oslo_concurrency.lockutils [req-6c97b3f7-fd19-425b-a42b-48f788087889 req-ba2f515f-15d9-4e33-8a85-025e40be019f service nova] Acquiring lock "d060a9f7-bd96-4e95-8780-1617a6ca7443-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.569526] env[61906]: DEBUG oslo_concurrency.lockutils [req-6c97b3f7-fd19-425b-a42b-48f788087889 req-ba2f515f-15d9-4e33-8a85-025e40be019f service nova] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.569696] env[61906]: DEBUG oslo_concurrency.lockutils [req-6c97b3f7-fd19-425b-a42b-48f788087889 req-ba2f515f-15d9-4e33-8a85-025e40be019f service nova] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.569870] env[61906]: DEBUG nova.compute.manager [req-6c97b3f7-fd19-425b-a42b-48f788087889 req-ba2f515f-15d9-4e33-8a85-025e40be019f service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] No waiting events found dispatching network-vif-plugged-e4fb390e-8245-4c57-a3b3-ef4556435df6 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 892.570178] env[61906]: WARNING nova.compute.manager [req-6c97b3f7-fd19-425b-a42b-48f788087889 req-ba2f515f-15d9-4e33-8a85-025e40be019f service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Received unexpected event network-vif-plugged-e4fb390e-8245-4c57-a3b3-ef4556435df6 for instance with vm_state building and task_state spawning. [ 892.764675] env[61906]: DEBUG nova.compute.utils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.766527] env[61906]: DEBUG nova.objects.instance [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lazy-loading 'numa_topology' on Instance uuid 33e288d5-9065-4606-b17a-c4c90bcbd533 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.767732] env[61906]: DEBUG nova.compute.manager [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 892.767934] env[61906]: DEBUG nova.network.neutron [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 892.779242] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "d0792521-4637-4ce7-a579-71caa80ba38c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.407s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.809091] env[61906]: DEBUG nova.policy [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ec22db431334e7886acb497ebfc7eb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ac9a6686e7184fb1913e84a77985b449', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 892.837152] env[61906]: DEBUG nova.network.neutron [-] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.861606] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f63c65fc-c28f-4149-b89b-61383b97e123 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.871370] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd23093-1a35-4369-9221-0aaca34825af {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.901898] env[61906]: DEBUG nova.compute.manager [req-63b33a54-2c32-48f2-abe0-83f3c01d5c51 req-cba6f7d8-6e6d-469b-8be7-2a2379d032f1 service nova] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Detach interface failed, port_id=ba127004-dc19-4fc4-b84b-97584ac68f34, reason: Instance 943c57f9-74c6-4465-9773-3fb01e78127b could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 892.926089] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356816, 'name': Rename_Task, 'duration_secs': 0.87801} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.926549] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 892.926639] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75d3f2f5-232d-45d7-a19a-73712a01d716 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.933757] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 892.933757] env[61906]: value = "task-1356818" [ 892.933757] env[61906]: _type = "Task" [ 892.933757] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.942437] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356818, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.978307] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356817, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.133100] env[61906]: DEBUG nova.network.neutron [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Successfully updated port: e4fb390e-8245-4c57-a3b3-ef4556435df6 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 893.192636] env[61906]: DEBUG nova.network.neutron [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Successfully created port: 2f799c25-3b8b-4759-aadf-35fc1e98022e {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 893.269121] env[61906]: DEBUG nova.compute.manager [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 893.272502] env[61906]: DEBUG nova.objects.base [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Object Instance<33e288d5-9065-4606-b17a-c4c90bcbd533> lazy-loaded attributes: resources,numa_topology {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 893.340778] env[61906]: INFO nova.compute.manager [-] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Took 1.22 seconds to deallocate network for instance. [ 893.446818] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356818, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.477890] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356817, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.647004} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.480939] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 99e5c55d-79ae-47a6-8500-79fc68291650/99e5c55d-79ae-47a6-8500-79fc68291650.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.481215] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.481692] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0dd89901-db1f-4878-8972-ec1a8e4d0c4a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.489423] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 893.489423] env[61906]: value = "task-1356819" [ 893.489423] env[61906]: _type = "Task" [ 893.489423] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.494315] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129467a2-8547-4cde-8d69-7eda98f845ed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.501885] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356819, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.504999] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a13c46f-7feb-4021-81da-360db35886ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.537419] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214954dd-b8c5-478e-a2d1-5344f323c934 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.545115] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc438896-5934-4d45-9e5e-003ab6d2a53a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.558996] env[61906]: DEBUG nova.compute.provider_tree [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.635966] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "refresh_cache-d060a9f7-bd96-4e95-8780-1617a6ca7443" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.635966] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquired lock "refresh_cache-d060a9f7-bd96-4e95-8780-1617a6ca7443" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.636118] env[61906]: DEBUG nova.network.neutron [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 893.848603] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.947788] env[61906]: DEBUG oslo_vmware.api [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356818, 'name': PowerOnVM_Task, 'duration_secs': 0.531336} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.948271] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 893.948364] env[61906]: INFO nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Took 8.86 seconds to spawn the instance on the hypervisor. [ 893.948534] env[61906]: DEBUG nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 893.949339] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b00335ef-3f2f-408b-8f09-efd0e584e195 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.000577] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356819, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068821} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.000783] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.001565] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9092a42-62d0-4504-9dd2-3abc797d5095 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.025850] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 99e5c55d-79ae-47a6-8500-79fc68291650/99e5c55d-79ae-47a6-8500-79fc68291650.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.026258] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efc1c686-8bd2-4617-a4eb-b8efa7860565 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.049830] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 894.049830] env[61906]: value = "task-1356820" [ 894.049830] env[61906]: _type = "Task" [ 894.049830] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.060085] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356820, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.065719] env[61906]: DEBUG nova.scheduler.client.report [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.179900] env[61906]: DEBUG nova.network.neutron [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 894.281314] env[61906]: DEBUG nova.compute.manager [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 894.317763] env[61906]: DEBUG nova.virt.hardware [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='7bb44d51c285c1562e0d88dc8179dc22',container_format='bare',created_at=2024-10-24T14:10:56Z,direct_url=,disk_format='vmdk',id=cc54d929-47af-46bb-8c53-a38fb410a7cc,min_disk=1,min_ram=0,name='tempest-test-snap-2126557259',owner='ac9a6686e7184fb1913e84a77985b449',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-24T14:11:12Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 894.318140] env[61906]: DEBUG nova.virt.hardware [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 894.318385] env[61906]: DEBUG nova.virt.hardware [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.318531] env[61906]: DEBUG nova.virt.hardware [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 894.318687] env[61906]: DEBUG nova.virt.hardware [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.318893] env[61906]: DEBUG nova.virt.hardware [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 894.319066] env[61906]: DEBUG nova.virt.hardware [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 894.319410] env[61906]: DEBUG nova.virt.hardware [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 894.319894] env[61906]: DEBUG nova.virt.hardware [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 894.320150] env[61906]: DEBUG nova.virt.hardware [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 894.320481] env[61906]: DEBUG nova.virt.hardware [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 894.321814] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa83999-2bb7-4c69-800f-86122480edd3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.333883] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc47481-8e52-40a2-948c-8277faba17a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.400945] env[61906]: DEBUG nova.network.neutron [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Updating instance_info_cache with network_info: [{"id": "e4fb390e-8245-4c57-a3b3-ef4556435df6", "address": "fa:16:3e:c1:3c:b6", "network": {"id": "b4ea38e3-d0e0-48a3-9632-772ca8257e37", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-540734152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17447d31611a46bbabae82a3ea2a2f22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90328c7b-15c4-4742-805b-755248d67029", "external-id": "nsx-vlan-transportzone-860", "segmentation_id": 860, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4fb390e-82", "ovs_interfaceid": "e4fb390e-8245-4c57-a3b3-ef4556435df6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.467750] env[61906]: INFO nova.compute.manager [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Took 18.05 seconds to build instance. [ 894.473665] env[61906]: DEBUG nova.compute.manager [req-5e8f6ad0-27aa-4488-b67e-618a3c6a9946 req-2d050c70-2271-4c3b-bfcb-41d642c11409 service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Received event network-changed-e4fb390e-8245-4c57-a3b3-ef4556435df6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.473665] env[61906]: DEBUG nova.compute.manager [req-5e8f6ad0-27aa-4488-b67e-618a3c6a9946 req-2d050c70-2271-4c3b-bfcb-41d642c11409 service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Refreshing instance network info cache due to event network-changed-e4fb390e-8245-4c57-a3b3-ef4556435df6. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 894.473665] env[61906]: DEBUG oslo_concurrency.lockutils [req-5e8f6ad0-27aa-4488-b67e-618a3c6a9946 req-2d050c70-2271-4c3b-bfcb-41d642c11409 service nova] Acquiring lock "refresh_cache-d060a9f7-bd96-4e95-8780-1617a6ca7443" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.561701] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356820, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.571439] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.311s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.574059] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.527s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.575576] env[61906]: INFO nova.compute.claims [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 894.725802] env[61906]: DEBUG nova.compute.manager [req-ad7bed5a-272f-4220-b60b-a0285d37b6b2 req-0b226ae0-ac9f-4063-9731-c73540d4a564 service nova] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Received event network-vif-plugged-2f799c25-3b8b-4759-aadf-35fc1e98022e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.726052] env[61906]: DEBUG oslo_concurrency.lockutils [req-ad7bed5a-272f-4220-b60b-a0285d37b6b2 req-0b226ae0-ac9f-4063-9731-c73540d4a564 service nova] Acquiring lock "e35580b3-7177-46fd-bb86-b013efbf4911-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.726396] env[61906]: DEBUG oslo_concurrency.lockutils [req-ad7bed5a-272f-4220-b60b-a0285d37b6b2 req-0b226ae0-ac9f-4063-9731-c73540d4a564 service nova] Lock "e35580b3-7177-46fd-bb86-b013efbf4911-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.726461] env[61906]: DEBUG oslo_concurrency.lockutils [req-ad7bed5a-272f-4220-b60b-a0285d37b6b2 req-0b226ae0-ac9f-4063-9731-c73540d4a564 service nova] Lock "e35580b3-7177-46fd-bb86-b013efbf4911-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.726634] env[61906]: DEBUG nova.compute.manager [req-ad7bed5a-272f-4220-b60b-a0285d37b6b2 req-0b226ae0-ac9f-4063-9731-c73540d4a564 service nova] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] No waiting events found dispatching network-vif-plugged-2f799c25-3b8b-4759-aadf-35fc1e98022e {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 894.726784] env[61906]: WARNING nova.compute.manager [req-ad7bed5a-272f-4220-b60b-a0285d37b6b2 req-0b226ae0-ac9f-4063-9731-c73540d4a564 service nova] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Received unexpected event network-vif-plugged-2f799c25-3b8b-4759-aadf-35fc1e98022e for instance with vm_state building and task_state spawning. [ 894.825197] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "d0792521-4637-4ce7-a579-71caa80ba38c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.825502] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "d0792521-4637-4ce7-a579-71caa80ba38c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.825902] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "d0792521-4637-4ce7-a579-71caa80ba38c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.826161] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "d0792521-4637-4ce7-a579-71caa80ba38c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.826348] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "d0792521-4637-4ce7-a579-71caa80ba38c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.828663] env[61906]: INFO nova.compute.manager [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Terminating instance [ 894.830664] env[61906]: DEBUG nova.compute.manager [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 894.830772] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 894.831590] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be36d291-78f9-4fff-b529-da657bef87c7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.839585] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.839744] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c637f8c0-eb65-4023-8532-9ea6a9cb54c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.845964] env[61906]: DEBUG oslo_vmware.api [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 894.845964] env[61906]: value = "task-1356821" [ 894.845964] env[61906]: _type = "Task" [ 894.845964] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.853839] env[61906]: DEBUG oslo_vmware.api [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356821, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.903692] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Releasing lock "refresh_cache-d060a9f7-bd96-4e95-8780-1617a6ca7443" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.904175] env[61906]: DEBUG nova.compute.manager [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Instance network_info: |[{"id": "e4fb390e-8245-4c57-a3b3-ef4556435df6", "address": "fa:16:3e:c1:3c:b6", "network": {"id": "b4ea38e3-d0e0-48a3-9632-772ca8257e37", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-540734152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17447d31611a46bbabae82a3ea2a2f22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90328c7b-15c4-4742-805b-755248d67029", "external-id": "nsx-vlan-transportzone-860", "segmentation_id": 860, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4fb390e-82", "ovs_interfaceid": "e4fb390e-8245-4c57-a3b3-ef4556435df6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 894.904503] env[61906]: DEBUG oslo_concurrency.lockutils [req-5e8f6ad0-27aa-4488-b67e-618a3c6a9946 req-2d050c70-2271-4c3b-bfcb-41d642c11409 service nova] Acquired lock "refresh_cache-d060a9f7-bd96-4e95-8780-1617a6ca7443" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.904698] env[61906]: DEBUG nova.network.neutron [req-5e8f6ad0-27aa-4488-b67e-618a3c6a9946 req-2d050c70-2271-4c3b-bfcb-41d642c11409 service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Refreshing network info cache for port e4fb390e-8245-4c57-a3b3-ef4556435df6 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 894.906094] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c1:3c:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '90328c7b-15c4-4742-805b-755248d67029', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e4fb390e-8245-4c57-a3b3-ef4556435df6', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.914395] env[61906]: DEBUG oslo.service.loopingcall [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.915447] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 894.915708] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3785c838-0882-492f-b22d-b4943c858bb6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.937950] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.937950] env[61906]: value = "task-1356822" [ 894.937950] env[61906]: _type = "Task" [ 894.937950] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.945777] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356822, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.970326] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47bdf58-03ef-4697-b34d-799f77786acc tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "03bd64e4-16fc-4659-a428-a2d9e7205b81" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.560s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.986588] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "03bd64e4-16fc-4659-a428-a2d9e7205b81" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.987041] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "03bd64e4-16fc-4659-a428-a2d9e7205b81" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.987353] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "03bd64e4-16fc-4659-a428-a2d9e7205b81-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.987645] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "03bd64e4-16fc-4659-a428-a2d9e7205b81-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.987851] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "03bd64e4-16fc-4659-a428-a2d9e7205b81-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.990553] env[61906]: INFO nova.compute.manager [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Terminating instance [ 894.992990] env[61906]: DEBUG nova.compute.manager [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 894.993247] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 894.994245] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1354dab-fd2b-4571-9805-fa4ddcedbe84 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.002907] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.003244] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-efc7ba64-554a-47a9-a5ea-ba98d631dc38 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.009596] env[61906]: DEBUG oslo_vmware.api [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 895.009596] env[61906]: value = "task-1356823" [ 895.009596] env[61906]: _type = "Task" [ 895.009596] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.018059] env[61906]: DEBUG oslo_vmware.api [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356823, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.059231] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356820, 'name': ReconfigVM_Task, 'duration_secs': 0.523367} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.059575] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 99e5c55d-79ae-47a6-8500-79fc68291650/99e5c55d-79ae-47a6-8500-79fc68291650.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 895.060260] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0d00e97-9921-42cc-97a1-c36f41b1bf31 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.066709] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 895.066709] env[61906]: value = "task-1356824" [ 895.066709] env[61906]: _type = "Task" [ 895.066709] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.078233] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356824, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.085973] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aae2ed29-57c9-4119-89b8-d6cb7c7b18cc tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "33e288d5-9065-4606-b17a-c4c90bcbd533" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 27.010s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.085973] env[61906]: DEBUG oslo_concurrency.lockutils [None req-babcfaa3-0806-4d41-bc6d-297a421247d2 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "33e288d5-9065-4606-b17a-c4c90bcbd533" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 4.823s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.085973] env[61906]: DEBUG oslo_concurrency.lockutils [None req-babcfaa3-0806-4d41-bc6d-297a421247d2 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "33e288d5-9065-4606-b17a-c4c90bcbd533-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.086239] env[61906]: DEBUG oslo_concurrency.lockutils [None req-babcfaa3-0806-4d41-bc6d-297a421247d2 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "33e288d5-9065-4606-b17a-c4c90bcbd533-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.086400] env[61906]: DEBUG oslo_concurrency.lockutils [None req-babcfaa3-0806-4d41-bc6d-297a421247d2 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "33e288d5-9065-4606-b17a-c4c90bcbd533-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.089788] env[61906]: INFO nova.compute.manager [None req-babcfaa3-0806-4d41-bc6d-297a421247d2 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Terminating instance [ 895.091726] env[61906]: DEBUG nova.compute.manager [None req-babcfaa3-0806-4d41-bc6d-297a421247d2 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 895.091927] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-babcfaa3-0806-4d41-bc6d-297a421247d2 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 895.092216] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cffe37fc-0eb2-4533-a4d6-70a4f5d8510f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.102327] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194c86ae-58e6-476d-a5fa-06c1ee76611d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.135188] env[61906]: WARNING nova.virt.vmwareapi.vmops [None req-babcfaa3-0806-4d41-bc6d-297a421247d2 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 33e288d5-9065-4606-b17a-c4c90bcbd533 could not be found. [ 895.135552] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-babcfaa3-0806-4d41-bc6d-297a421247d2 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 895.136073] env[61906]: INFO nova.compute.manager [None req-babcfaa3-0806-4d41-bc6d-297a421247d2 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Took 0.04 seconds to destroy the instance on the hypervisor. [ 895.136428] env[61906]: DEBUG oslo.service.loopingcall [None req-babcfaa3-0806-4d41-bc6d-297a421247d2 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.136865] env[61906]: DEBUG nova.compute.manager [-] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 895.136996] env[61906]: DEBUG nova.network.neutron [-] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 895.307445] env[61906]: DEBUG nova.network.neutron [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Successfully updated port: 2f799c25-3b8b-4759-aadf-35fc1e98022e {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.356694] env[61906]: DEBUG oslo_vmware.api [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356821, 'name': PowerOffVM_Task, 'duration_secs': 0.20077} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.357000] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.357233] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 895.357492] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-375192a3-0eb3-4a35-b04a-5b175e3b4cdc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.420205] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.420459] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.420644] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Deleting the datastore file [datastore2] d0792521-4637-4ce7-a579-71caa80ba38c {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.421227] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b70d89ff-bf78-4342-8bd3-17ad4491df09 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.428209] env[61906]: DEBUG oslo_vmware.api [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 895.428209] env[61906]: value = "task-1356826" [ 895.428209] env[61906]: _type = "Task" [ 895.428209] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.436963] env[61906]: DEBUG oslo_vmware.api [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356826, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.445621] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356822, 'name': CreateVM_Task, 'duration_secs': 0.428514} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.448016] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 895.448760] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.448970] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.449327] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.449943] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a28c4753-9662-4c0b-af35-140bea7c588d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.454871] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 895.454871] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522d494c-ec30-720d-ddf4-f28604c75f48" [ 895.454871] env[61906]: _type = "Task" [ 895.454871] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.463248] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522d494c-ec30-720d-ddf4-f28604c75f48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.519675] env[61906]: DEBUG oslo_vmware.api [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356823, 'name': PowerOffVM_Task, 'duration_secs': 0.217558} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.519985] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.520245] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 895.520459] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-68fe52ba-48f2-4386-8874-2522e7de84ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.577547] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356824, 'name': Rename_Task, 'duration_secs': 0.173341} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.578363] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 895.578363] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ca4f223-363e-4e6f-9593-732b7d4a9b9c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.587459] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 895.587459] env[61906]: value = "task-1356828" [ 895.587459] env[61906]: _type = "Task" [ 895.587459] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.594960] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.595257] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.595483] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Deleting the datastore file [datastore2] 03bd64e4-16fc-4659-a428-a2d9e7205b81 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.595800] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66281f90-f2f1-42e3-a25c-d23e3c3796e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.605040] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356828, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.605040] env[61906]: DEBUG oslo_vmware.api [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for the task: (returnval){ [ 895.605040] env[61906]: value = "task-1356829" [ 895.605040] env[61906]: _type = "Task" [ 895.605040] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.612820] env[61906]: DEBUG oslo_vmware.api [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356829, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.678926] env[61906]: DEBUG nova.network.neutron [req-5e8f6ad0-27aa-4488-b67e-618a3c6a9946 req-2d050c70-2271-4c3b-bfcb-41d642c11409 service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Updated VIF entry in instance network info cache for port e4fb390e-8245-4c57-a3b3-ef4556435df6. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 895.679417] env[61906]: DEBUG nova.network.neutron [req-5e8f6ad0-27aa-4488-b67e-618a3c6a9946 req-2d050c70-2271-4c3b-bfcb-41d642c11409 service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Updating instance_info_cache with network_info: [{"id": "e4fb390e-8245-4c57-a3b3-ef4556435df6", "address": "fa:16:3e:c1:3c:b6", "network": {"id": "b4ea38e3-d0e0-48a3-9632-772ca8257e37", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-540734152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17447d31611a46bbabae82a3ea2a2f22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90328c7b-15c4-4742-805b-755248d67029", "external-id": "nsx-vlan-transportzone-860", "segmentation_id": 860, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4fb390e-82", "ovs_interfaceid": "e4fb390e-8245-4c57-a3b3-ef4556435df6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.796549] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46c5be4-cd52-4b3c-80bf-ebe0ffcb0050 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.806538] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b335f1ab-b766-4255-beb0-1bab809271ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.810873] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "refresh_cache-e35580b3-7177-46fd-bb86-b013efbf4911" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.811135] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "refresh_cache-e35580b3-7177-46fd-bb86-b013efbf4911" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.811370] env[61906]: DEBUG nova.network.neutron [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 895.846268] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b9da83-b780-452f-a7f5-bd5344ee7c1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.855058] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0b2afd-7b3c-4586-b760-cf47d5e4e312 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.869410] env[61906]: DEBUG nova.compute.provider_tree [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.898930] env[61906]: DEBUG nova.network.neutron [-] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.937889] env[61906]: DEBUG oslo_vmware.api [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356826, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201827} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.938118] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.938322] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 895.938511] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 895.938696] env[61906]: INFO nova.compute.manager [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 895.938947] env[61906]: DEBUG oslo.service.loopingcall [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.939176] env[61906]: DEBUG nova.compute.manager [-] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 895.939277] env[61906]: DEBUG nova.network.neutron [-] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 895.965229] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522d494c-ec30-720d-ddf4-f28604c75f48, 'name': SearchDatastore_Task, 'duration_secs': 0.013557} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.965565] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.965809] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.966784] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.966784] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.966784] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.966784] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-48111910-2d34-45f4-9865-480ee6ebee6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.976017] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.976289] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 895.978986] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0f007ff-1332-4bed-b6d1-76f3244becd6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.984909] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 895.984909] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527c3322-36f0-7ae2-e2dc-9f9d2347ffbb" [ 895.984909] env[61906]: _type = "Task" [ 895.984909] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.992542] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527c3322-36f0-7ae2-e2dc-9f9d2347ffbb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.100213] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356828, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.114074] env[61906]: DEBUG oslo_vmware.api [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Task: {'id': task-1356829, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204541} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.114313] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 896.115029] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 896.115029] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 896.115029] env[61906]: INFO nova.compute.manager [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Took 1.12 seconds to destroy the instance on the hypervisor. [ 896.115240] env[61906]: DEBUG oslo.service.loopingcall [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.115469] env[61906]: DEBUG nova.compute.manager [-] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 896.115591] env[61906]: DEBUG nova.network.neutron [-] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 896.184296] env[61906]: DEBUG oslo_concurrency.lockutils [req-5e8f6ad0-27aa-4488-b67e-618a3c6a9946 req-2d050c70-2271-4c3b-bfcb-41d642c11409 service nova] Releasing lock "refresh_cache-d060a9f7-bd96-4e95-8780-1617a6ca7443" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.364651] env[61906]: DEBUG nova.network.neutron [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 896.376749] env[61906]: DEBUG nova.scheduler.client.report [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.401498] env[61906]: INFO nova.compute.manager [-] [instance: 33e288d5-9065-4606-b17a-c4c90bcbd533] Took 1.26 seconds to deallocate network for instance. [ 896.496223] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527c3322-36f0-7ae2-e2dc-9f9d2347ffbb, 'name': SearchDatastore_Task, 'duration_secs': 0.011572} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.497097] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44ca8c2e-208a-4098-afef-ef0e52be0073 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.503985] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 896.503985] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5277b815-5152-89c2-fe49-b9dc602ebbb1" [ 896.503985] env[61906]: _type = "Task" [ 896.503985] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.510221] env[61906]: DEBUG nova.compute.manager [req-30eb315f-c3c2-48f5-a128-03c6c69edf1f req-99ad2ae6-3354-4fbf-83a2-f58c8f749eb9 service nova] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Received event network-changed-2f799c25-3b8b-4759-aadf-35fc1e98022e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.510804] env[61906]: DEBUG nova.compute.manager [req-30eb315f-c3c2-48f5-a128-03c6c69edf1f req-99ad2ae6-3354-4fbf-83a2-f58c8f749eb9 service nova] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Refreshing instance network info cache due to event network-changed-2f799c25-3b8b-4759-aadf-35fc1e98022e. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.511222] env[61906]: DEBUG oslo_concurrency.lockutils [req-30eb315f-c3c2-48f5-a128-03c6c69edf1f req-99ad2ae6-3354-4fbf-83a2-f58c8f749eb9 service nova] Acquiring lock "refresh_cache-e35580b3-7177-46fd-bb86-b013efbf4911" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.516720] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5277b815-5152-89c2-fe49-b9dc602ebbb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.576745] env[61906]: DEBUG nova.network.neutron [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Updating instance_info_cache with network_info: [{"id": "2f799c25-3b8b-4759-aadf-35fc1e98022e", "address": "fa:16:3e:66:0c:73", "network": {"id": "5b29c9e0-23c9-42ee-9be4-60178bddedbe", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643906254-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac9a6686e7184fb1913e84a77985b449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f799c25-3b", "ovs_interfaceid": "2f799c25-3b8b-4759-aadf-35fc1e98022e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.601379] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356828, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.704886] env[61906]: DEBUG nova.network.neutron [-] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.882532] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.883139] env[61906]: DEBUG nova.compute.manager [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 896.886049] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.037s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.886173] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.913659] env[61906]: INFO nova.scheduler.client.report [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Deleted allocations for instance 943c57f9-74c6-4465-9773-3fb01e78127b [ 897.002474] env[61906]: DEBUG nova.network.neutron [-] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.015381] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5277b815-5152-89c2-fe49-b9dc602ebbb1, 'name': SearchDatastore_Task, 'duration_secs': 0.014677} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.016326] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.016598] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] d060a9f7-bd96-4e95-8780-1617a6ca7443/d060a9f7-bd96-4e95-8780-1617a6ca7443.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 897.017027] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8dc4a01-400a-413b-b3e7-3a73d5108bc2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.024100] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 897.024100] env[61906]: value = "task-1356830" [ 897.024100] env[61906]: _type = "Task" [ 897.024100] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.032447] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356830, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.080176] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "refresh_cache-e35580b3-7177-46fd-bb86-b013efbf4911" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.080523] env[61906]: DEBUG nova.compute.manager [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Instance network_info: |[{"id": "2f799c25-3b8b-4759-aadf-35fc1e98022e", "address": "fa:16:3e:66:0c:73", "network": {"id": "5b29c9e0-23c9-42ee-9be4-60178bddedbe", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643906254-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac9a6686e7184fb1913e84a77985b449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f799c25-3b", "ovs_interfaceid": "2f799c25-3b8b-4759-aadf-35fc1e98022e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 897.081012] env[61906]: DEBUG oslo_concurrency.lockutils [req-30eb315f-c3c2-48f5-a128-03c6c69edf1f req-99ad2ae6-3354-4fbf-83a2-f58c8f749eb9 service nova] Acquired lock "refresh_cache-e35580b3-7177-46fd-bb86-b013efbf4911" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.081269] env[61906]: DEBUG nova.network.neutron [req-30eb315f-c3c2-48f5-a128-03c6c69edf1f req-99ad2ae6-3354-4fbf-83a2-f58c8f749eb9 service nova] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Refreshing network info cache for port 2f799c25-3b8b-4759-aadf-35fc1e98022e {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 897.082618] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:0c:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c02dd284-ab80-451c-93eb-48c8360acb9c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2f799c25-3b8b-4759-aadf-35fc1e98022e', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.091360] env[61906]: DEBUG oslo.service.loopingcall [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.095089] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 897.099025] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-684a7861-938a-43e6-96e4-29554abf2c04 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.119964] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356828, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.121311] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.121311] env[61906]: value = "task-1356831" [ 897.121311] env[61906]: _type = "Task" [ 897.121311] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.129077] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356831, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.207714] env[61906]: INFO nova.compute.manager [-] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Took 1.27 seconds to deallocate network for instance. [ 897.355407] env[61906]: DEBUG nova.network.neutron [req-30eb315f-c3c2-48f5-a128-03c6c69edf1f req-99ad2ae6-3354-4fbf-83a2-f58c8f749eb9 service nova] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Updated VIF entry in instance network info cache for port 2f799c25-3b8b-4759-aadf-35fc1e98022e. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 897.355794] env[61906]: DEBUG nova.network.neutron [req-30eb315f-c3c2-48f5-a128-03c6c69edf1f req-99ad2ae6-3354-4fbf-83a2-f58c8f749eb9 service nova] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Updating instance_info_cache with network_info: [{"id": "2f799c25-3b8b-4759-aadf-35fc1e98022e", "address": "fa:16:3e:66:0c:73", "network": {"id": "5b29c9e0-23c9-42ee-9be4-60178bddedbe", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643906254-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac9a6686e7184fb1913e84a77985b449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f799c25-3b", "ovs_interfaceid": "2f799c25-3b8b-4759-aadf-35fc1e98022e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.391151] env[61906]: DEBUG nova.compute.utils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 897.392599] env[61906]: DEBUG nova.compute.manager [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 897.392775] env[61906]: DEBUG nova.network.neutron [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 897.423416] env[61906]: DEBUG oslo_concurrency.lockutils [None req-3392b735-7b40-46ed-9199-8b7ff8e4477c tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "943c57f9-74c6-4465-9773-3fb01e78127b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.930s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.430956] env[61906]: DEBUG oslo_concurrency.lockutils [None req-babcfaa3-0806-4d41-bc6d-297a421247d2 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "33e288d5-9065-4606-b17a-c4c90bcbd533" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.345s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.439344] env[61906]: DEBUG nova.policy [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca80129d492e4d68b96ef14a63336ef3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb85e88fd4f54e0db2ff131f81137f64', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 897.504904] env[61906]: INFO nova.compute.manager [-] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Took 1.39 seconds to deallocate network for instance. [ 897.533967] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356830, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.585068] env[61906]: DEBUG oslo_vmware.rw_handles [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526c36c8-32c4-857d-1fd0-2b827cb61bf7/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 897.586115] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa71ccf7-0ffc-4e75-ae31-0cc70e8cd8f3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.593502] env[61906]: DEBUG oslo_vmware.rw_handles [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526c36c8-32c4-857d-1fd0-2b827cb61bf7/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 897.593679] env[61906]: ERROR oslo_vmware.rw_handles [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526c36c8-32c4-857d-1fd0-2b827cb61bf7/disk-0.vmdk due to incomplete transfer. [ 897.593913] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-96e00005-dbba-430a-885a-4c6b698a413c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.607215] env[61906]: DEBUG oslo_vmware.api [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356828, 'name': PowerOnVM_Task, 'duration_secs': 1.620693} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.607215] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 897.607311] env[61906]: INFO nova.compute.manager [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Took 10.20 seconds to spawn the instance on the hypervisor. [ 897.607690] env[61906]: DEBUG nova.compute.manager [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.607690] env[61906]: DEBUG oslo_vmware.rw_handles [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526c36c8-32c4-857d-1fd0-2b827cb61bf7/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 897.607897] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Uploaded image 4e467722-4f47-4b26-9f34-13132817ea0e to the Glance image server {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 897.610291] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 897.611115] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6342f92-5af4-4f59-93a0-8c18532abb97 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.615143] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e31cd200-d006-4c2e-9267-97293fb4b0de {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.624149] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 897.624149] env[61906]: value = "task-1356832" [ 897.624149] env[61906]: _type = "Task" [ 897.624149] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.636103] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356831, 'name': CreateVM_Task, 'duration_secs': 0.448406} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.639310] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 897.639689] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356832, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.641122] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cc54d929-47af-46bb-8c53-a38fb410a7cc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.641122] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cc54d929-47af-46bb-8c53-a38fb410a7cc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.641926] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cc54d929-47af-46bb-8c53-a38fb410a7cc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 897.641926] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62203cbd-adb1-4370-82c0-c26947d1040b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.647931] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 897.647931] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5213c4d1-ae6e-7848-acdc-46f4340ab154" [ 897.647931] env[61906]: _type = "Task" [ 897.647931] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.655052] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5213c4d1-ae6e-7848-acdc-46f4340ab154, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.718012] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.718012] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.718278] env[61906]: DEBUG nova.objects.instance [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lazy-loading 'resources' on Instance uuid d0792521-4637-4ce7-a579-71caa80ba38c {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.752265] env[61906]: DEBUG nova.network.neutron [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Successfully created port: c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 897.858819] env[61906]: DEBUG oslo_concurrency.lockutils [req-30eb315f-c3c2-48f5-a128-03c6c69edf1f req-99ad2ae6-3354-4fbf-83a2-f58c8f749eb9 service nova] Releasing lock "refresh_cache-e35580b3-7177-46fd-bb86-b013efbf4911" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.859139] env[61906]: DEBUG nova.compute.manager [req-30eb315f-c3c2-48f5-a128-03c6c69edf1f req-99ad2ae6-3354-4fbf-83a2-f58c8f749eb9 service nova] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Received event network-vif-deleted-6664c3db-45ba-4a24-a4a1-2b00df77b08f {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.859366] env[61906]: INFO nova.compute.manager [req-30eb315f-c3c2-48f5-a128-03c6c69edf1f req-99ad2ae6-3354-4fbf-83a2-f58c8f749eb9 service nova] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Neutron deleted interface 6664c3db-45ba-4a24-a4a1-2b00df77b08f; detaching it from the instance and deleting it from the info cache [ 897.859551] env[61906]: DEBUG nova.network.neutron [req-30eb315f-c3c2-48f5-a128-03c6c69edf1f req-99ad2ae6-3354-4fbf-83a2-f58c8f749eb9 service nova] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.895829] env[61906]: DEBUG nova.compute.manager [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 898.011709] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.035467] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356830, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.84313} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.035746] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] d060a9f7-bd96-4e95-8780-1617a6ca7443/d060a9f7-bd96-4e95-8780-1617a6ca7443.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 898.035970] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.036390] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5a555709-8733-4ff9-8836-47a108670343 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.046582] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 898.046582] env[61906]: value = "task-1356833" [ 898.046582] env[61906]: _type = "Task" [ 898.046582] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.052573] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356833, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.140234] env[61906]: INFO nova.compute.manager [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Took 19.72 seconds to build instance. [ 898.149942] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356832, 'name': Destroy_Task, 'duration_secs': 0.523735} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.153521] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Destroyed the VM [ 898.153788] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 898.154068] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e558862f-49cd-4e64-9fe6-5630fee46b5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.162823] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cc54d929-47af-46bb-8c53-a38fb410a7cc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.163145] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Processing image cc54d929-47af-46bb-8c53-a38fb410a7cc {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.163442] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cc54d929-47af-46bb-8c53-a38fb410a7cc/cc54d929-47af-46bb-8c53-a38fb410a7cc.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.163635] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cc54d929-47af-46bb-8c53-a38fb410a7cc/cc54d929-47af-46bb-8c53-a38fb410a7cc.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.163865] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.164213] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 898.164213] env[61906]: value = "task-1356834" [ 898.164213] env[61906]: _type = "Task" [ 898.164213] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.164695] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d4592b9d-09b7-44b7-8d0b-8ef7d4f21f55 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.175237] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356834, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.178053] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.178291] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.179188] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c83b613a-815e-4b21-b1d3-f901ce8ebd1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.184770] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 898.184770] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527fff36-772e-9f94-66b5-a861be192469" [ 898.184770] env[61906]: _type = "Task" [ 898.184770] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.192788] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527fff36-772e-9f94-66b5-a861be192469, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.364774] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a97bf295-21c7-4cbc-9091-cec7123a06b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.377507] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56fd018e-5cef-4b74-b9e7-7032d06a7502 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.405106] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-122d5826-cc31-44a6-9bbe-108c2a055ca2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.416468] env[61906]: DEBUG nova.compute.manager [req-30eb315f-c3c2-48f5-a128-03c6c69edf1f req-99ad2ae6-3354-4fbf-83a2-f58c8f749eb9 service nova] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Detach interface failed, port_id=6664c3db-45ba-4a24-a4a1-2b00df77b08f, reason: Instance d0792521-4637-4ce7-a579-71caa80ba38c could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 898.421575] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d98928-5eec-4fcb-8994-96b728cfdfff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.456206] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a11b171-f8c6-45cd-aa94-b8115e90b5f5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.464774] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188c2496-e512-4b0c-be1d-e558a3b3961f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.479716] env[61906]: DEBUG nova.compute.provider_tree [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.537310] env[61906]: DEBUG nova.compute.manager [req-b0050081-dbce-481f-9a00-3a785c44fdba req-569a99cd-79fb-431f-ae6a-25fe7625bc2a service nova] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Received event network-vif-deleted-cc178483-0a7d-4037-8bd3-5d0a6a4ffefb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.554463] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356833, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09533} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.554832] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.555932] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0e3917-98c0-4abf-884e-bd39e4949d24 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.578680] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] d060a9f7-bd96-4e95-8780-1617a6ca7443/d060a9f7-bd96-4e95-8780-1617a6ca7443.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.579393] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac591de4-3820-4dc8-82fb-4c40252a6c7f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.599020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.599020] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.602129] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 898.602129] env[61906]: value = "task-1356835" [ 898.602129] env[61906]: _type = "Task" [ 898.602129] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.610327] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356835, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.642140] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c8cb4c55-32aa-4030-87b7-1ab7286d3249 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "99e5c55d-79ae-47a6-8500-79fc68291650" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.232s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.677414] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356834, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.695657] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Preparing fetch location {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 898.696053] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Fetch image to [datastore1] OSTACK_IMG_ba8e07c5-8b71-4cbe-872f-fcd1f0770cfd/OSTACK_IMG_ba8e07c5-8b71-4cbe-872f-fcd1f0770cfd.vmdk {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 898.696331] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Downloading stream optimized image cc54d929-47af-46bb-8c53-a38fb410a7cc to [datastore1] OSTACK_IMG_ba8e07c5-8b71-4cbe-872f-fcd1f0770cfd/OSTACK_IMG_ba8e07c5-8b71-4cbe-872f-fcd1f0770cfd.vmdk on the data store datastore1 as vApp {{(pid=61906) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 898.696603] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Downloading image file data cc54d929-47af-46bb-8c53-a38fb410a7cc to the ESX as VM named 'OSTACK_IMG_ba8e07c5-8b71-4cbe-872f-fcd1f0770cfd' {{(pid=61906) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 898.781404] env[61906]: DEBUG oslo_vmware.rw_handles [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 898.781404] env[61906]: value = "resgroup-9" [ 898.781404] env[61906]: _type = "ResourcePool" [ 898.781404] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 898.781720] env[61906]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-8afeb910-d9c7-462b-b8ee-fed82e90b51f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.803881] env[61906]: DEBUG oslo_vmware.rw_handles [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lease: (returnval){ [ 898.803881] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5293c532-f6f6-b6e4-b896-3506a3714a73" [ 898.803881] env[61906]: _type = "HttpNfcLease" [ 898.803881] env[61906]: } obtained for vApp import into resource pool (val){ [ 898.803881] env[61906]: value = "resgroup-9" [ 898.803881] env[61906]: _type = "ResourcePool" [ 898.803881] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 898.804257] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the lease: (returnval){ [ 898.804257] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5293c532-f6f6-b6e4-b896-3506a3714a73" [ 898.804257] env[61906]: _type = "HttpNfcLease" [ 898.804257] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 898.810155] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 898.810155] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5293c532-f6f6-b6e4-b896-3506a3714a73" [ 898.810155] env[61906]: _type = "HttpNfcLease" [ 898.810155] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 898.910036] env[61906]: DEBUG nova.compute.manager [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 898.934060] env[61906]: DEBUG nova.virt.hardware [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.934060] env[61906]: DEBUG nova.virt.hardware [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.934060] env[61906]: DEBUG nova.virt.hardware [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.934060] env[61906]: DEBUG nova.virt.hardware [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.934060] env[61906]: DEBUG nova.virt.hardware [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.934060] env[61906]: DEBUG nova.virt.hardware [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.934444] env[61906]: DEBUG nova.virt.hardware [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.934444] env[61906]: DEBUG nova.virt.hardware [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.934623] env[61906]: DEBUG nova.virt.hardware [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.934807] env[61906]: DEBUG nova.virt.hardware [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.934987] env[61906]: DEBUG nova.virt.hardware [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.936036] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437ca1b7-4a28-4b11-a083-c8329c054c22 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.949315] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0edb3f2-8609-4a28-903d-c1e62fb58024 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.983099] env[61906]: DEBUG nova.scheduler.client.report [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.101462] env[61906]: DEBUG nova.compute.manager [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 899.113227] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356835, 'name': ReconfigVM_Task, 'duration_secs': 0.281474} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.113523] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Reconfigured VM instance instance-00000056 to attach disk [datastore1] d060a9f7-bd96-4e95-8780-1617a6ca7443/d060a9f7-bd96-4e95-8780-1617a6ca7443.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 899.114958] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ba9353d4-167d-441f-b09a-2916bc3db84e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.121369] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 899.121369] env[61906]: value = "task-1356837" [ 899.121369] env[61906]: _type = "Task" [ 899.121369] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.133780] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356837, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.177490] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356834, 'name': RemoveSnapshot_Task, 'duration_secs': 0.54979} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.177763] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 899.178217] env[61906]: DEBUG nova.compute.manager [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.178998] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55988d11-d572-4a09-9ef2-7fd9f00396ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.314721] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 899.314721] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5293c532-f6f6-b6e4-b896-3506a3714a73" [ 899.314721] env[61906]: _type = "HttpNfcLease" [ 899.314721] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 899.489986] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.770s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.490950] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.479s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.491433] env[61906]: DEBUG nova.objects.instance [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lazy-loading 'resources' on Instance uuid 03bd64e4-16fc-4659-a428-a2d9e7205b81 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.515641] env[61906]: INFO nova.scheduler.client.report [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Deleted allocations for instance d0792521-4637-4ce7-a579-71caa80ba38c [ 899.577316] env[61906]: DEBUG nova.network.neutron [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Successfully updated port: c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 899.626448] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.632193] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356837, 'name': Rename_Task, 'duration_secs': 0.23071} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.634181] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.634181] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ce076d1-780c-4291-ab1d-37e65c146a81 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.639544] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 899.639544] env[61906]: value = "task-1356838" [ 899.639544] env[61906]: _type = "Task" [ 899.639544] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.647596] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356838, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.692162] env[61906]: INFO nova.compute.manager [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Shelve offloading [ 899.693856] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.694140] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-869cb8db-f17a-4ebd-abf0-29e0348c2865 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.700494] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 899.700494] env[61906]: value = "task-1356839" [ 899.700494] env[61906]: _type = "Task" [ 899.700494] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.710118] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356839, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.812910] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 899.812910] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5293c532-f6f6-b6e4-b896-3506a3714a73" [ 899.812910] env[61906]: _type = "HttpNfcLease" [ 899.812910] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 899.813270] env[61906]: DEBUG oslo_vmware.rw_handles [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 899.813270] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5293c532-f6f6-b6e4-b896-3506a3714a73" [ 899.813270] env[61906]: _type = "HttpNfcLease" [ 899.813270] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 899.814046] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e75d578a-8b8b-4a2e-be0d-384adc26b1d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.821146] env[61906]: DEBUG oslo_vmware.rw_handles [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520ecf78-3dcc-539d-e516-7aae50ab638b/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 899.821330] env[61906]: DEBUG oslo_vmware.rw_handles [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520ecf78-3dcc-539d-e516-7aae50ab638b/disk-0.vmdk. {{(pid=61906) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 899.885687] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-51a1304c-b58e-4a2a-aa5f-cbee7e473d1c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.025874] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2c02ece-f024-419b-bbdb-097b1f442df3 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "d0792521-4637-4ce7-a579-71caa80ba38c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.200s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.078952] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "refresh_cache-bc773559-fe14-485f-8bdc-1d4aab186521" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.079136] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "refresh_cache-bc773559-fe14-485f-8bdc-1d4aab186521" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.079316] env[61906]: DEBUG nova.network.neutron [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.153322] env[61906]: DEBUG oslo_vmware.api [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356838, 'name': PowerOnVM_Task, 'duration_secs': 0.4609} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.158263] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 900.158491] env[61906]: INFO nova.compute.manager [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Took 8.17 seconds to spawn the instance on the hypervisor. [ 900.158702] env[61906]: DEBUG nova.compute.manager [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.159790] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e8a7d4-5323-48dc-bdf4-3b4d3892b66d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.174034] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d302e24f-e368-4f8c-812a-2b1690e7a66c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.182274] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027c12c3-3ad6-4443-accb-5d1357926f2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.223348] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f41bbe9-0494-48d6-b8a2-2941477856ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.239008] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] VM already powered off {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 900.239422] env[61906]: DEBUG nova.compute.manager [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.240818] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e8a150-f86c-49c5-a5b7-084eb5f81ee3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.245534] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b693df06-ac0e-4059-9ea3-bd9761be2f62 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.259162] env[61906]: DEBUG oslo_concurrency.lockutils [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.259505] env[61906]: DEBUG oslo_concurrency.lockutils [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.259760] env[61906]: DEBUG nova.network.neutron [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.277120] env[61906]: DEBUG nova.compute.provider_tree [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.565590] env[61906]: DEBUG nova.compute.manager [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Received event network-changed-9a08ec03-408f-4d57-842d-3f68f53dcb29 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.565763] env[61906]: DEBUG nova.compute.manager [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Refreshing instance network info cache due to event network-changed-9a08ec03-408f-4d57-842d-3f68f53dcb29. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 900.565968] env[61906]: DEBUG oslo_concurrency.lockutils [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] Acquiring lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.566211] env[61906]: DEBUG oslo_concurrency.lockutils [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] Acquired lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.566361] env[61906]: DEBUG nova.network.neutron [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Refreshing network info cache for port 9a08ec03-408f-4d57-842d-3f68f53dcb29 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 900.580065] env[61906]: DEBUG nova.network.neutron [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updating instance_info_cache with network_info: [{"id": "f119e802-0a46-4390-927d-805c2e23a4b4", "address": "fa:16:3e:2c:b5:5c", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119e802-0a", "ovs_interfaceid": "f119e802-0a46-4390-927d-805c2e23a4b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.613460] env[61906]: DEBUG nova.network.neutron [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 900.692221] env[61906]: INFO nova.compute.manager [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Took 13.07 seconds to build instance. [ 900.764723] env[61906]: DEBUG nova.network.neutron [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Updating instance_info_cache with network_info: [{"id": "c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2", "address": "fa:16:3e:95:19:be", "network": {"id": "9e0705ba-7b9f-46fb-8603-01598a888212", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-705593104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb85e88fd4f54e0db2ff131f81137f64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc83d72a2-80", "ovs_interfaceid": "c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.776695] env[61906]: DEBUG oslo_vmware.rw_handles [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Completed reading data from the image iterator. {{(pid=61906) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 900.777559] env[61906]: DEBUG oslo_vmware.rw_handles [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520ecf78-3dcc-539d-e516-7aae50ab638b/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 900.779729] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf1ad60-31a0-42ff-b545-02ac1eff73b7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.784157] env[61906]: DEBUG nova.scheduler.client.report [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 900.789362] env[61906]: DEBUG oslo_vmware.rw_handles [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520ecf78-3dcc-539d-e516-7aae50ab638b/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 900.789539] env[61906]: DEBUG oslo_vmware.rw_handles [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520ecf78-3dcc-539d-e516-7aae50ab638b/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 900.789992] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-abbca542-98df-4d18-9362-7adea1c39023 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.085257] env[61906]: DEBUG oslo_concurrency.lockutils [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.194931] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7c29af67-236e-4388-ae48-931362445ab7 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.586s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.268999] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "refresh_cache-bc773559-fe14-485f-8bdc-1d4aab186521" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.269387] env[61906]: DEBUG nova.compute.manager [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Instance network_info: |[{"id": "c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2", "address": "fa:16:3e:95:19:be", "network": {"id": "9e0705ba-7b9f-46fb-8603-01598a888212", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-705593104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb85e88fd4f54e0db2ff131f81137f64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc83d72a2-80", "ovs_interfaceid": "c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 901.269899] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:19:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4c5eb94-841c-4713-985a-8fc4117fbaf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.277506] env[61906]: DEBUG oslo.service.loopingcall [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.277606] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 901.277822] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8e9c853-dd31-486f-b963-0fa15442caa7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.292608] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.802s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.294692] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.668s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.296318] env[61906]: INFO nova.compute.claims [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.303994] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.303994] env[61906]: value = "task-1356840" [ 901.303994] env[61906]: _type = "Task" [ 901.303994] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.314795] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356840, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.321738] env[61906]: INFO nova.scheduler.client.report [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Deleted allocations for instance 03bd64e4-16fc-4659-a428-a2d9e7205b81 [ 901.353964] env[61906]: DEBUG oslo_vmware.rw_handles [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520ecf78-3dcc-539d-e516-7aae50ab638b/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 901.354214] env[61906]: INFO nova.virt.vmwareapi.images [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Downloaded image file data cc54d929-47af-46bb-8c53-a38fb410a7cc [ 901.355156] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8687fbe0-4699-49d7-a3ad-3547659c6d95 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.374066] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a624a96e-d2e7-404e-941b-3245a4db4530 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.403223] env[61906]: INFO nova.virt.vmwareapi.images [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] The imported VM was unregistered [ 901.408022] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Caching image {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 901.408022] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Creating directory with path [datastore1] devstack-image-cache_base/cc54d929-47af-46bb-8c53-a38fb410a7cc {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.408022] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ddd4531-696c-4a11-b4d5-dc780304cdfa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.426358] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Created directory with path [datastore1] devstack-image-cache_base/cc54d929-47af-46bb-8c53-a38fb410a7cc {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.426596] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_ba8e07c5-8b71-4cbe-872f-fcd1f0770cfd/OSTACK_IMG_ba8e07c5-8b71-4cbe-872f-fcd1f0770cfd.vmdk to [datastore1] devstack-image-cache_base/cc54d929-47af-46bb-8c53-a38fb410a7cc/cc54d929-47af-46bb-8c53-a38fb410a7cc.vmdk. {{(pid=61906) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 901.426923] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-45ab5447-52f3-414b-87ca-ec2937d49a0d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.435780] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 901.435780] env[61906]: value = "task-1356842" [ 901.435780] env[61906]: _type = "Task" [ 901.435780] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.443896] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356842, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.821818] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356840, 'name': CreateVM_Task, 'duration_secs': 0.346733} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.822011] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.822801] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.823007] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.823452] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.824010] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e318a4c0-f1cd-4fcf-a55e-9349cf7d1451 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.828583] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dd7d759e-d5e2-41c8-ba31-87e12a3d4842 tempest-MultipleCreateTestJSON-994386374 tempest-MultipleCreateTestJSON-994386374-project-member] Lock "03bd64e4-16fc-4659-a428-a2d9e7205b81" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.842s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.839015] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 901.839015] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c82931-7c19-1644-7d14-fac521cc490c" [ 901.839015] env[61906]: _type = "Task" [ 901.839015] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.849632] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c82931-7c19-1644-7d14-fac521cc490c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.901688] env[61906]: DEBUG nova.network.neutron [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updated VIF entry in instance network info cache for port 9a08ec03-408f-4d57-842d-3f68f53dcb29. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.902015] env[61906]: DEBUG nova.network.neutron [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updating instance_info_cache with network_info: [{"id": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "address": "fa:16:3e:b8:da:a6", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a08ec03-40", "ovs_interfaceid": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.950251] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356842, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.127078] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 902.128166] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4f80c6-863c-41a0-b099-cd9578166415 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.137211] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 902.137506] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-384d81e5-8e46-4159-a4ef-8f81f7cd8428 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.223600] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 902.223976] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 902.224070] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleting the datastore file [datastore2] cb879dc0-3af7-4279-aa28-66e2b8b4286d {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.224320] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-86071e29-b0cd-4d30-9ad8-d8f786a3fbcd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.231624] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 902.231624] env[61906]: value = "task-1356844" [ 902.231624] env[61906]: _type = "Task" [ 902.231624] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.241721] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356844, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.361881] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c82931-7c19-1644-7d14-fac521cc490c, 'name': SearchDatastore_Task, 'duration_secs': 0.090497} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.362313] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.362786] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.363060] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.363434] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.363631] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.364103] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a079d401-d01a-4af2-8397-0693b2cf2f2a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.387952] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.388253] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.389030] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d380f3be-c095-4abb-812d-ebdc1a8951bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.396843] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 902.396843] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c0ab47-2d38-d678-ae81-132e1f4d5166" [ 902.396843] env[61906]: _type = "Task" [ 902.396843] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.405041] env[61906]: DEBUG oslo_concurrency.lockutils [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] Releasing lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.405320] env[61906]: DEBUG nova.compute.manager [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Received event network-vif-plugged-c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.405513] env[61906]: DEBUG oslo_concurrency.lockutils [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] Acquiring lock "bc773559-fe14-485f-8bdc-1d4aab186521-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.405869] env[61906]: DEBUG oslo_concurrency.lockutils [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] Lock "bc773559-fe14-485f-8bdc-1d4aab186521-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.405869] env[61906]: DEBUG oslo_concurrency.lockutils [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] Lock "bc773559-fe14-485f-8bdc-1d4aab186521-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.406057] env[61906]: DEBUG nova.compute.manager [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] No waiting events found dispatching network-vif-plugged-c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.406243] env[61906]: WARNING nova.compute.manager [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Received unexpected event network-vif-plugged-c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2 for instance with vm_state building and task_state spawning. [ 902.406736] env[61906]: DEBUG nova.compute.manager [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Received event network-changed-c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.406736] env[61906]: DEBUG nova.compute.manager [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Refreshing instance network info cache due to event network-changed-c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.406842] env[61906]: DEBUG oslo_concurrency.lockutils [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] Acquiring lock "refresh_cache-bc773559-fe14-485f-8bdc-1d4aab186521" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.406929] env[61906]: DEBUG oslo_concurrency.lockutils [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] Acquired lock "refresh_cache-bc773559-fe14-485f-8bdc-1d4aab186521" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.407139] env[61906]: DEBUG nova.network.neutron [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Refreshing network info cache for port c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 902.408302] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c0ab47-2d38-d678-ae81-132e1f4d5166, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.447865] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356842, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.499339] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f861e4e-6739-45bf-a41c-0e8c564737d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.508015] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51dee329-b800-458d-9e9e-f0b8de786fec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.539668] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc0193f-d4a8-4fcd-ae39-c187e112c92e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.548296] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32708e24-d63b-4f6f-80dc-69838dc26531 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.567788] env[61906]: DEBUG nova.compute.provider_tree [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.611222] env[61906]: DEBUG nova.compute.manager [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Received event network-vif-unplugged-f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.611556] env[61906]: DEBUG oslo_concurrency.lockutils [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] Acquiring lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.611794] env[61906]: DEBUG oslo_concurrency.lockutils [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.611997] env[61906]: DEBUG oslo_concurrency.lockutils [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.612192] env[61906]: DEBUG nova.compute.manager [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] No waiting events found dispatching network-vif-unplugged-f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.612369] env[61906]: WARNING nova.compute.manager [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Received unexpected event network-vif-unplugged-f119e802-0a46-4390-927d-805c2e23a4b4 for instance with vm_state shelved and task_state shelving_offloading. [ 902.612553] env[61906]: DEBUG nova.compute.manager [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Received event network-changed-e4fb390e-8245-4c57-a3b3-ef4556435df6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.612713] env[61906]: DEBUG nova.compute.manager [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Refreshing instance network info cache due to event network-changed-e4fb390e-8245-4c57-a3b3-ef4556435df6. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.612905] env[61906]: DEBUG oslo_concurrency.lockutils [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] Acquiring lock "refresh_cache-d060a9f7-bd96-4e95-8780-1617a6ca7443" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.613104] env[61906]: DEBUG oslo_concurrency.lockutils [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] Acquired lock "refresh_cache-d060a9f7-bd96-4e95-8780-1617a6ca7443" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.613280] env[61906]: DEBUG nova.network.neutron [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Refreshing network info cache for port e4fb390e-8245-4c57-a3b3-ef4556435df6 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 902.742965] env[61906]: DEBUG oslo_vmware.api [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356844, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.206298} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.743288] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.743483] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 902.743655] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 902.772747] env[61906]: INFO nova.scheduler.client.report [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleted allocations for instance cb879dc0-3af7-4279-aa28-66e2b8b4286d [ 902.916992] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c0ab47-2d38-d678-ae81-132e1f4d5166, 'name': SearchDatastore_Task, 'duration_secs': 0.129404} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.918306] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b578c2d-5c8b-4440-bda1-6f29c292839d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.924150] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 902.924150] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c8eefe-cba3-f12d-0993-c9f54d04dcf6" [ 902.924150] env[61906]: _type = "Task" [ 902.924150] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.932964] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c8eefe-cba3-f12d-0993-c9f54d04dcf6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.948555] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356842, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.071588] env[61906]: DEBUG nova.scheduler.client.report [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.210361] env[61906]: DEBUG nova.network.neutron [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Updated VIF entry in instance network info cache for port c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 903.210810] env[61906]: DEBUG nova.network.neutron [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Updating instance_info_cache with network_info: [{"id": "c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2", "address": "fa:16:3e:95:19:be", "network": {"id": "9e0705ba-7b9f-46fb-8603-01598a888212", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-705593104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb85e88fd4f54e0db2ff131f81137f64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc83d72a2-80", "ovs_interfaceid": "c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.277498] env[61906]: DEBUG oslo_concurrency.lockutils [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.426606] env[61906]: DEBUG nova.network.neutron [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Updated VIF entry in instance network info cache for port e4fb390e-8245-4c57-a3b3-ef4556435df6. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 903.427059] env[61906]: DEBUG nova.network.neutron [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Updating instance_info_cache with network_info: [{"id": "e4fb390e-8245-4c57-a3b3-ef4556435df6", "address": "fa:16:3e:c1:3c:b6", "network": {"id": "b4ea38e3-d0e0-48a3-9632-772ca8257e37", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-540734152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.134", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17447d31611a46bbabae82a3ea2a2f22", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "90328c7b-15c4-4742-805b-755248d67029", "external-id": "nsx-vlan-transportzone-860", "segmentation_id": 860, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4fb390e-82", "ovs_interfaceid": "e4fb390e-8245-4c57-a3b3-ef4556435df6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.439674] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c8eefe-cba3-f12d-0993-c9f54d04dcf6, 'name': SearchDatastore_Task, 'duration_secs': 0.110229} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.442692] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.442967] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] bc773559-fe14-485f-8bdc-1d4aab186521/bc773559-fe14-485f-8bdc-1d4aab186521.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.443814] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9814fc24-66a8-44cb-afd6-b7837bc39355 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.450520] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356842, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.451687] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 903.451687] env[61906]: value = "task-1356845" [ 903.451687] env[61906]: _type = "Task" [ 903.451687] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.460118] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.576626] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.282s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.577247] env[61906]: DEBUG nova.compute.manager [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 903.580108] env[61906]: DEBUG oslo_concurrency.lockutils [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.303s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.580356] env[61906]: DEBUG nova.objects.instance [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lazy-loading 'resources' on Instance uuid cb879dc0-3af7-4279-aa28-66e2b8b4286d {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.714751] env[61906]: DEBUG oslo_concurrency.lockutils [req-861dfe47-581d-4e1d-a07d-476ebbbab2a6 req-8fa6555f-4845-45d4-b6fa-dea256d59314 service nova] Releasing lock "refresh_cache-bc773559-fe14-485f-8bdc-1d4aab186521" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.934882] env[61906]: DEBUG oslo_concurrency.lockutils [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] Releasing lock "refresh_cache-d060a9f7-bd96-4e95-8780-1617a6ca7443" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.935198] env[61906]: DEBUG nova.compute.manager [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Received event network-changed-f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.935328] env[61906]: DEBUG nova.compute.manager [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Refreshing instance network info cache due to event network-changed-f119e802-0a46-4390-927d-805c2e23a4b4. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 903.935561] env[61906]: DEBUG oslo_concurrency.lockutils [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] Acquiring lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.935710] env[61906]: DEBUG oslo_concurrency.lockutils [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] Acquired lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.935908] env[61906]: DEBUG nova.network.neutron [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Refreshing network info cache for port f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 903.950311] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356842, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.962928] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.083055] env[61906]: DEBUG nova.objects.instance [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lazy-loading 'numa_topology' on Instance uuid cb879dc0-3af7-4279-aa28-66e2b8b4286d {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.085095] env[61906]: DEBUG nova.compute.utils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.087253] env[61906]: DEBUG nova.compute.manager [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 904.087253] env[61906]: DEBUG nova.network.neutron [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 904.149021] env[61906]: DEBUG nova.policy [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf4fea47b929461785f4a487a56977a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3093466de87a45a2ac673c25bea19bc3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.455303] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356842, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.460310] env[61906]: DEBUG nova.network.neutron [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Successfully created port: 505e72fa-25c4-4cc1-981f-d34507bb4b97 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 904.467822] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.587610] env[61906]: DEBUG nova.objects.base [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 904.591252] env[61906]: DEBUG nova.compute.manager [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 904.773058] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e72545-4e2f-484b-b1a3-d139f458197f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.784698] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d832f58-566c-43fe-b4c0-2600474318e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.820024] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e602da-15ea-4c0d-ada2-456da95172fb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.829311] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b2e246-24e9-4860-8540-c2a233eaae65 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.846046] env[61906]: DEBUG nova.compute.provider_tree [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.953880] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356842, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.045366} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.954282] env[61906]: INFO nova.virt.vmwareapi.ds_util [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_ba8e07c5-8b71-4cbe-872f-fcd1f0770cfd/OSTACK_IMG_ba8e07c5-8b71-4cbe-872f-fcd1f0770cfd.vmdk to [datastore1] devstack-image-cache_base/cc54d929-47af-46bb-8c53-a38fb410a7cc/cc54d929-47af-46bb-8c53-a38fb410a7cc.vmdk. [ 904.954533] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Cleaning up location [datastore1] OSTACK_IMG_ba8e07c5-8b71-4cbe-872f-fcd1f0770cfd {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 904.954778] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_ba8e07c5-8b71-4cbe-872f-fcd1f0770cfd {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 904.955054] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc2c4341-0c6a-46ef-980d-f55e910d9b30 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.965852] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356845, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.455728} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.968584] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] bc773559-fe14-485f-8bdc-1d4aab186521/bc773559-fe14-485f-8bdc-1d4aab186521.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 904.968584] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.968584] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 904.968584] env[61906]: value = "task-1356846" [ 904.968584] env[61906]: _type = "Task" [ 904.968584] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.968584] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0e1a7845-0284-435f-88f3-aed40a7393da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.977599] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356846, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.978972] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 904.978972] env[61906]: value = "task-1356847" [ 904.978972] env[61906]: _type = "Task" [ 904.978972] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.982804] env[61906]: DEBUG nova.network.neutron [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updated VIF entry in instance network info cache for port f119e802-0a46-4390-927d-805c2e23a4b4. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 904.983171] env[61906]: DEBUG nova.network.neutron [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updating instance_info_cache with network_info: [{"id": "f119e802-0a46-4390-927d-805c2e23a4b4", "address": "fa:16:3e:2c:b5:5c", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapf119e802-0a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.989663] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356847, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.185497] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.349434] env[61906]: DEBUG nova.scheduler.client.report [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.446588] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.446847] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.479855] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356846, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.082594} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.483044] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 905.483228] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cc54d929-47af-46bb-8c53-a38fb410a7cc/cc54d929-47af-46bb-8c53-a38fb410a7cc.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.483512] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cc54d929-47af-46bb-8c53-a38fb410a7cc/cc54d929-47af-46bb-8c53-a38fb410a7cc.vmdk to [datastore1] e35580b3-7177-46fd-bb86-b013efbf4911/e35580b3-7177-46fd-bb86-b013efbf4911.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 905.483768] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-154a89c9-d56d-4dda-bfdf-001bba68f195 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.485956] env[61906]: DEBUG oslo_concurrency.lockutils [req-9e19ee8e-3cb8-428d-b186-bceb57f64076 req-ebd761d9-96ff-455c-bf07-5bca7cd8621c service nova] Releasing lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.490095] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356847, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.31006} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.491199] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 905.491532] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 905.491532] env[61906]: value = "task-1356848" [ 905.491532] env[61906]: _type = "Task" [ 905.491532] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.492177] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6708aa1-ba88-49f1-8d22-0545c71cc9ba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.515709] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] bc773559-fe14-485f-8bdc-1d4aab186521/bc773559-fe14-485f-8bdc-1d4aab186521.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.518847] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76619720-639d-4a63-9869-b61b25f063bf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.532880] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356848, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.537939] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 905.537939] env[61906]: value = "task-1356849" [ 905.537939] env[61906]: _type = "Task" [ 905.537939] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.544933] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356849, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.603259] env[61906]: DEBUG nova.compute.manager [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 905.628931] env[61906]: DEBUG nova.virt.hardware [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 905.629209] env[61906]: DEBUG nova.virt.hardware [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 905.629384] env[61906]: DEBUG nova.virt.hardware [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.629567] env[61906]: DEBUG nova.virt.hardware [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 905.629749] env[61906]: DEBUG nova.virt.hardware [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.629918] env[61906]: DEBUG nova.virt.hardware [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 905.630163] env[61906]: DEBUG nova.virt.hardware [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 905.630333] env[61906]: DEBUG nova.virt.hardware [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 905.630508] env[61906]: DEBUG nova.virt.hardware [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 905.630679] env[61906]: DEBUG nova.virt.hardware [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 905.630884] env[61906]: DEBUG nova.virt.hardware [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.631748] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-182dd5e6-85e1-4e5b-9e8d-cde22d8a3a8f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.639360] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd50e29-686d-4d3c-ae91-429f1ae82430 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.854149] env[61906]: DEBUG oslo_concurrency.lockutils [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.274s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.950175] env[61906]: DEBUG nova.compute.utils [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 906.003883] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356848, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.048100] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356849, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.214688] env[61906]: DEBUG nova.compute.manager [req-60c57300-5ace-4a00-9ea2-a5aacf651bb7 req-0f4984f1-e6d6-4ada-b8cf-81f947708fd6 service nova] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Received event network-vif-plugged-505e72fa-25c4-4cc1-981f-d34507bb4b97 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.214688] env[61906]: DEBUG oslo_concurrency.lockutils [req-60c57300-5ace-4a00-9ea2-a5aacf651bb7 req-0f4984f1-e6d6-4ada-b8cf-81f947708fd6 service nova] Acquiring lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.214688] env[61906]: DEBUG oslo_concurrency.lockutils [req-60c57300-5ace-4a00-9ea2-a5aacf651bb7 req-0f4984f1-e6d6-4ada-b8cf-81f947708fd6 service nova] Lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.214853] env[61906]: DEBUG oslo_concurrency.lockutils [req-60c57300-5ace-4a00-9ea2-a5aacf651bb7 req-0f4984f1-e6d6-4ada-b8cf-81f947708fd6 service nova] Lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.215205] env[61906]: DEBUG nova.compute.manager [req-60c57300-5ace-4a00-9ea2-a5aacf651bb7 req-0f4984f1-e6d6-4ada-b8cf-81f947708fd6 service nova] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] No waiting events found dispatching network-vif-plugged-505e72fa-25c4-4cc1-981f-d34507bb4b97 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.215520] env[61906]: WARNING nova.compute.manager [req-60c57300-5ace-4a00-9ea2-a5aacf651bb7 req-0f4984f1-e6d6-4ada-b8cf-81f947708fd6 service nova] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Received unexpected event network-vif-plugged-505e72fa-25c4-4cc1-981f-d34507bb4b97 for instance with vm_state building and task_state spawning. [ 906.280657] env[61906]: DEBUG nova.network.neutron [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Successfully updated port: 505e72fa-25c4-4cc1-981f-d34507bb4b97 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.362754] env[61906]: DEBUG oslo_concurrency.lockutils [None req-507b2a57-d89b-43a1-bb6c-c5eedcd2e0e6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 22.174s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.363908] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.178s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.364412] env[61906]: INFO nova.compute.manager [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Unshelving [ 906.452585] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.504433] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356848, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.547073] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356849, 'name': ReconfigVM_Task, 'duration_secs': 0.567022} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.547389] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Reconfigured VM instance instance-00000058 to attach disk [datastore1] bc773559-fe14-485f-8bdc-1d4aab186521/bc773559-fe14-485f-8bdc-1d4aab186521.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.548058] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-efce89b8-e42a-4e0d-b545-38b91b142e79 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.553846] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 906.553846] env[61906]: value = "task-1356850" [ 906.553846] env[61906]: _type = "Task" [ 906.553846] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.561121] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356850, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.784149] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "refresh_cache-a4c9f463-98e7-49eb-8c01-c8f0ee98a884" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.784321] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "refresh_cache-a4c9f463-98e7-49eb-8c01-c8f0ee98a884" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.784465] env[61906]: DEBUG nova.network.neutron [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.006985] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356848, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.063792] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356850, 'name': Rename_Task, 'duration_secs': 0.441503} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.064095] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 907.064339] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-68478969-dd96-43a8-973f-511d4e467b90 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.070121] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 907.070121] env[61906]: value = "task-1356851" [ 907.070121] env[61906]: _type = "Task" [ 907.070121] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.077491] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356851, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.313598] env[61906]: DEBUG nova.network.neutron [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 907.374174] env[61906]: DEBUG nova.compute.utils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 907.455653] env[61906]: DEBUG nova.network.neutron [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Updating instance_info_cache with network_info: [{"id": "505e72fa-25c4-4cc1-981f-d34507bb4b97", "address": "fa:16:3e:b2:d2:fa", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap505e72fa-25", "ovs_interfaceid": "505e72fa-25c4-4cc1-981f-d34507bb4b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.507727] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356848, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.517704] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.517978] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.518260] env[61906]: INFO nova.compute.manager [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Attaching volume cfc83076-1277-4f42-9c90-ef2e9aec2b8a to /dev/sdb [ 907.556751] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7375571b-124c-4e04-884c-89ee591968cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.565452] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbef331-9131-4f77-8205-06b33ad1d1d5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.579601] env[61906]: DEBUG nova.virt.block_device [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updating existing volume attachment record: 6654e18d-aa48-4a33-a0bb-a125684944fb {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 907.588581] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356851, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.877956] env[61906]: INFO nova.virt.block_device [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Booting with volume 4d5078bb-42a4-4eaa-b253-4e94f1321e51 at /dev/sdb [ 907.919181] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33c1076a-1a53-4aa1-9879-fba9923a1618 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.929657] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f3e0e5-7671-4b91-bbcc-485d34568bb1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.960516] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "refresh_cache-a4c9f463-98e7-49eb-8c01-c8f0ee98a884" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.960859] env[61906]: DEBUG nova.compute.manager [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Instance network_info: |[{"id": "505e72fa-25c4-4cc1-981f-d34507bb4b97", "address": "fa:16:3e:b2:d2:fa", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap505e72fa-25", "ovs_interfaceid": "505e72fa-25c4-4cc1-981f-d34507bb4b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 907.961543] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b2:d2:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30f1dacf-8988-41b8-aa8f-e9530f65ef46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '505e72fa-25c4-4cc1-981f-d34507bb4b97', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 907.969042] env[61906]: DEBUG oslo.service.loopingcall [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.969295] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4228156d-85ef-4514-8dab-3f824daa73e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.971416] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 907.971657] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9bd67c0b-c5e4-47e0-be6b-9fbd77d93b88 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.994871] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ebcc82f-3e7a-4705-a772-86203f2338b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.009112] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.009112] env[61906]: value = "task-1356855" [ 908.009112] env[61906]: _type = "Task" [ 908.009112] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.021264] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356855, 'name': CreateVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.024682] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356848, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.033615] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e57a9a-d262-4001-9bce-6881716c9e11 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.041848] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa642c05-c38f-4a52-9547-474b4d4cee25 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.057807] env[61906]: DEBUG nova.virt.block_device [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updating existing volume attachment record: 7b4d0a0d-6cf1-48d8-b2a1-67c7f8cbc6a0 {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 908.082366] env[61906]: DEBUG oslo_vmware.api [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356851, 'name': PowerOnVM_Task, 'duration_secs': 0.874067} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.082664] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 908.082901] env[61906]: INFO nova.compute.manager [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Took 9.17 seconds to spawn the instance on the hypervisor. [ 908.083104] env[61906]: DEBUG nova.compute.manager [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.083917] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d9b91f-39d3-4261-8c18-f1bd32869bb3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.248625] env[61906]: DEBUG nova.compute.manager [req-fbf21730-c449-497a-8397-5dfa9f82907e req-8cfa88f8-0ea6-4ced-b150-c6d5355b1400 service nova] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Received event network-changed-505e72fa-25c4-4cc1-981f-d34507bb4b97 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.248863] env[61906]: DEBUG nova.compute.manager [req-fbf21730-c449-497a-8397-5dfa9f82907e req-8cfa88f8-0ea6-4ced-b150-c6d5355b1400 service nova] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Refreshing instance network info cache due to event network-changed-505e72fa-25c4-4cc1-981f-d34507bb4b97. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 908.249080] env[61906]: DEBUG oslo_concurrency.lockutils [req-fbf21730-c449-497a-8397-5dfa9f82907e req-8cfa88f8-0ea6-4ced-b150-c6d5355b1400 service nova] Acquiring lock "refresh_cache-a4c9f463-98e7-49eb-8c01-c8f0ee98a884" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.249447] env[61906]: DEBUG oslo_concurrency.lockutils [req-fbf21730-c449-497a-8397-5dfa9f82907e req-8cfa88f8-0ea6-4ced-b150-c6d5355b1400 service nova] Acquired lock "refresh_cache-a4c9f463-98e7-49eb-8c01-c8f0ee98a884" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.249517] env[61906]: DEBUG nova.network.neutron [req-fbf21730-c449-497a-8397-5dfa9f82907e req-8cfa88f8-0ea6-4ced-b150-c6d5355b1400 service nova] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Refreshing network info cache for port 505e72fa-25c4-4cc1-981f-d34507bb4b97 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 908.524865] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356848, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.528251] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356855, 'name': CreateVM_Task, 'duration_secs': 0.370552} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.528491] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 908.529281] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.529503] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.529851] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.530147] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ce46e60-81d6-4b92-929a-2e753fe399a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.536712] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 908.536712] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5281f02b-2b87-720b-67e8-6377cc188ae8" [ 908.536712] env[61906]: _type = "Task" [ 908.536712] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.546765] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5281f02b-2b87-720b-67e8-6377cc188ae8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.605617] env[61906]: INFO nova.compute.manager [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Took 16.59 seconds to build instance. [ 909.026919] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356848, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.051787] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5281f02b-2b87-720b-67e8-6377cc188ae8, 'name': SearchDatastore_Task, 'duration_secs': 0.014689} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.052320] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.052733] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.053161] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.053441] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.053919] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.054374] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be9b0a80-29dc-40ff-acc2-107bfbb114b1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.062455] env[61906]: DEBUG nova.network.neutron [req-fbf21730-c449-497a-8397-5dfa9f82907e req-8cfa88f8-0ea6-4ced-b150-c6d5355b1400 service nova] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Updated VIF entry in instance network info cache for port 505e72fa-25c4-4cc1-981f-d34507bb4b97. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 909.062974] env[61906]: DEBUG nova.network.neutron [req-fbf21730-c449-497a-8397-5dfa9f82907e req-8cfa88f8-0ea6-4ced-b150-c6d5355b1400 service nova] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Updating instance_info_cache with network_info: [{"id": "505e72fa-25c4-4cc1-981f-d34507bb4b97", "address": "fa:16:3e:b2:d2:fa", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap505e72fa-25", "ovs_interfaceid": "505e72fa-25c4-4cc1-981f-d34507bb4b97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.066617] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.066723] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 909.068385] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35e12e23-7931-443b-8a46-d6cdeb6813b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.077908] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 909.077908] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fa4eda-c1dc-495a-583f-0dece31b10ab" [ 909.077908] env[61906]: _type = "Task" [ 909.077908] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.093245] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fa4eda-c1dc-495a-583f-0dece31b10ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.107802] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2fdc495c-84d9-4a1c-bd9b-b94cbffe6903 tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "bc773559-fe14-485f-8bdc-1d4aab186521" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.102s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.522061] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356848, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.948333} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.522240] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cc54d929-47af-46bb-8c53-a38fb410a7cc/cc54d929-47af-46bb-8c53-a38fb410a7cc.vmdk to [datastore1] e35580b3-7177-46fd-bb86-b013efbf4911/e35580b3-7177-46fd-bb86-b013efbf4911.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 909.522915] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53ff8e5-9e8c-4a28-8700-ccadfa6e62e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.545907] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] e35580b3-7177-46fd-bb86-b013efbf4911/e35580b3-7177-46fd-bb86-b013efbf4911.vmdk or device None with type streamOptimized {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.546430] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23f994b3-07a9-40aa-afe5-ca2c23635cdc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.565731] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 909.565731] env[61906]: value = "task-1356859" [ 909.565731] env[61906]: _type = "Task" [ 909.565731] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.568770] env[61906]: DEBUG oslo_concurrency.lockutils [req-fbf21730-c449-497a-8397-5dfa9f82907e req-8cfa88f8-0ea6-4ced-b150-c6d5355b1400 service nova] Releasing lock "refresh_cache-a4c9f463-98e7-49eb-8c01-c8f0ee98a884" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.575621] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356859, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.586559] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fa4eda-c1dc-495a-583f-0dece31b10ab, 'name': SearchDatastore_Task, 'duration_secs': 0.012185} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.587360] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18f90c31-3d47-46d6-9f88-bd1dcd0edd1c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.592509] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 909.592509] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d0d028-9327-df55-e096-9b13eee3c31c" [ 909.592509] env[61906]: _type = "Task" [ 909.592509] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.600930] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d0d028-9327-df55-e096-9b13eee3c31c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.692804] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 909.693206] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 909.693415] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 909.774729] env[61906]: DEBUG oslo_concurrency.lockutils [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "bc773559-fe14-485f-8bdc-1d4aab186521" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.775106] env[61906]: DEBUG oslo_concurrency.lockutils [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "bc773559-fe14-485f-8bdc-1d4aab186521" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.775332] env[61906]: DEBUG oslo_concurrency.lockutils [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "bc773559-fe14-485f-8bdc-1d4aab186521-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.775562] env[61906]: DEBUG oslo_concurrency.lockutils [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "bc773559-fe14-485f-8bdc-1d4aab186521-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.775786] env[61906]: DEBUG oslo_concurrency.lockutils [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "bc773559-fe14-485f-8bdc-1d4aab186521-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.778234] env[61906]: INFO nova.compute.manager [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Terminating instance [ 909.780221] env[61906]: DEBUG nova.compute.manager [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 909.780458] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 909.781356] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a041a2bf-8abf-46ad-904b-69acdbab00b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.789535] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.789781] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c24913f-57ec-44c8-8afd-63280dbac50e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.796506] env[61906]: DEBUG oslo_vmware.api [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 909.796506] env[61906]: value = "task-1356860" [ 909.796506] env[61906]: _type = "Task" [ 909.796506] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.804446] env[61906]: DEBUG oslo_vmware.api [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356860, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.076877] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356859, 'name': ReconfigVM_Task, 'duration_secs': 0.278564} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.077216] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Reconfigured VM instance instance-00000057 to attach disk [datastore1] e35580b3-7177-46fd-bb86-b013efbf4911/e35580b3-7177-46fd-bb86-b013efbf4911.vmdk or device None with type streamOptimized {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.077872] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4b2dbfbf-26c3-4dba-828a-ae056fa150d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.083445] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 910.083445] env[61906]: value = "task-1356862" [ 910.083445] env[61906]: _type = "Task" [ 910.083445] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.090957] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356862, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.100471] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d0d028-9327-df55-e096-9b13eee3c31c, 'name': SearchDatastore_Task, 'duration_secs': 0.010103} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.100666] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.100918] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] a4c9f463-98e7-49eb-8c01-c8f0ee98a884/a4c9f463-98e7-49eb-8c01-c8f0ee98a884.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 910.101166] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-201bac61-09aa-4314-a0fe-dae5af6b4f4c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.106838] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 910.106838] env[61906]: value = "task-1356863" [ 910.106838] env[61906]: _type = "Task" [ 910.106838] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.113710] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356863, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.306538] env[61906]: DEBUG oslo_vmware.api [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356860, 'name': PowerOffVM_Task, 'duration_secs': 0.258138} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.306870] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.307105] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 910.307408] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9cef82e-ab9d-4e23-89a0-3515ab780e5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.596730] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356862, 'name': Rename_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.617188] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 910.617404] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 910.617643] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Deleting the datastore file [datastore1] bc773559-fe14-485f-8bdc-1d4aab186521 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.620720] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-711db179-0caf-42f6-9066-9f6a4f73d3fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.622802] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356863, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.627839] env[61906]: DEBUG oslo_vmware.api [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for the task: (returnval){ [ 910.627839] env[61906]: value = "task-1356866" [ 910.627839] env[61906]: _type = "Task" [ 910.627839] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.637226] env[61906]: DEBUG oslo_vmware.api [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356866, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.094175] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356862, 'name': Rename_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.116805] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356863, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524661} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.117077] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] a4c9f463-98e7-49eb-8c01-c8f0ee98a884/a4c9f463-98e7-49eb-8c01-c8f0ee98a884.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 911.117330] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.117609] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e17a095-88a3-4e2b-8c1a-0295c26ca7fb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.124918] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 911.124918] env[61906]: value = "task-1356867" [ 911.124918] env[61906]: _type = "Task" [ 911.124918] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.133482] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356867, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.138066] env[61906]: DEBUG oslo_vmware.api [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Task: {'id': task-1356866, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147123} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.138337] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.138535] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 911.138717] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 911.138899] env[61906]: INFO nova.compute.manager [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Took 1.36 seconds to destroy the instance on the hypervisor. [ 911.139162] env[61906]: DEBUG oslo.service.loopingcall [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.139366] env[61906]: DEBUG nova.compute.manager [-] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.139492] env[61906]: DEBUG nova.network.neutron [-] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 911.390885] env[61906]: DEBUG nova.compute.manager [req-dd339c41-d99e-4265-99ee-cad8462a27ac req-1e4fc4fa-e387-47a0-8714-bb01876b3d1d service nova] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Received event network-vif-deleted-c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.391296] env[61906]: INFO nova.compute.manager [req-dd339c41-d99e-4265-99ee-cad8462a27ac req-1e4fc4fa-e387-47a0-8714-bb01876b3d1d service nova] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Neutron deleted interface c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2; detaching it from the instance and deleting it from the info cache [ 911.391296] env[61906]: DEBUG nova.network.neutron [req-dd339c41-d99e-4265-99ee-cad8462a27ac req-1e4fc4fa-e387-47a0-8714-bb01876b3d1d service nova] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.594898] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356862, 'name': Rename_Task, 'duration_secs': 1.147345} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.595041] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 911.595297] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9a2a8cc-e6b9-44ca-988d-13182cdd44ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.602140] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 911.602140] env[61906]: value = "task-1356868" [ 911.602140] env[61906]: _type = "Task" [ 911.602140] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.609626] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356868, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.633930] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356867, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070318} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.634234] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.635065] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700e4b36-d1de-4e00-880a-8641e324a64b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.656521] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] a4c9f463-98e7-49eb-8c01-c8f0ee98a884/a4c9f463-98e7-49eb-8c01-c8f0ee98a884.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.657183] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83e98894-48a0-47ce-a938-9e16d42a3924 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.676523] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 911.676523] env[61906]: value = "task-1356869" [ 911.676523] env[61906]: _type = "Task" [ 911.676523] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.684564] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356869, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.868688] env[61906]: DEBUG nova.network.neutron [-] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.894997] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0edfc76e-ae84-43b0-8d38-1d7ba7027825 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.904491] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a9561a-bb15-45f0-927d-bb873f8ddfaa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.934009] env[61906]: DEBUG nova.compute.manager [req-dd339c41-d99e-4265-99ee-cad8462a27ac req-1e4fc4fa-e387-47a0-8714-bb01876b3d1d service nova] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Detach interface failed, port_id=c83d72a2-805d-4f64-aa1c-fa11a0f2bfe2, reason: Instance bc773559-fe14-485f-8bdc-1d4aab186521 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 912.112559] env[61906]: DEBUG oslo_vmware.api [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356868, 'name': PowerOnVM_Task, 'duration_secs': 0.49348} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.112915] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 912.113128] env[61906]: INFO nova.compute.manager [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Took 17.83 seconds to spawn the instance on the hypervisor. [ 912.113308] env[61906]: DEBUG nova.compute.manager [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.114061] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b68f7c-ecbd-4c40-a859-228e73d5741a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.129872] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 912.130107] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-289013', 'volume_id': 'cfc83076-1277-4f42-9c90-ef2e9aec2b8a', 'name': 'volume-cfc83076-1277-4f42-9c90-ef2e9aec2b8a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63', 'attached_at': '', 'detached_at': '', 'volume_id': 'cfc83076-1277-4f42-9c90-ef2e9aec2b8a', 'serial': 'cfc83076-1277-4f42-9c90-ef2e9aec2b8a'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 912.130923] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef594cc3-6dc3-4d5e-b528-4e99f6c1de9f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.147045] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e117e4b2-ccd2-4abc-a994-9181962c99df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.172180] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] volume-cfc83076-1277-4f42-9c90-ef2e9aec2b8a/volume-cfc83076-1277-4f42-9c90-ef2e9aec2b8a.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.173172] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f1120df-6c97-4fc3-9302-87963c887519 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.194451] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356869, 'name': ReconfigVM_Task, 'duration_secs': 0.415952} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.195729] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Reconfigured VM instance instance-00000059 to attach disk [datastore2] a4c9f463-98e7-49eb-8c01-c8f0ee98a884/a4c9f463-98e7-49eb-8c01-c8f0ee98a884.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.196417] env[61906]: DEBUG oslo_vmware.api [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 912.196417] env[61906]: value = "task-1356870" [ 912.196417] env[61906]: _type = "Task" [ 912.196417] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.196611] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e7dcc7b-177e-4bb2-a989-fabea6beed6b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.205918] env[61906]: DEBUG oslo_vmware.api [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356870, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.206989] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 912.206989] env[61906]: value = "task-1356871" [ 912.206989] env[61906]: _type = "Task" [ 912.206989] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.219416] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356871, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.237886] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.238097] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquired lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.238285] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 912.371218] env[61906]: INFO nova.compute.manager [-] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Took 1.23 seconds to deallocate network for instance. [ 912.629745] env[61906]: INFO nova.compute.manager [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Took 23.96 seconds to build instance. [ 912.709587] env[61906]: DEBUG oslo_vmware.api [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356870, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.719165] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356871, 'name': Rename_Task, 'duration_secs': 0.164829} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.719511] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 912.719746] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2129548b-2c05-4ba4-9c67-d551878c2422 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.725489] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 912.725489] env[61906]: value = "task-1356872" [ 912.725489] env[61906]: _type = "Task" [ 912.725489] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.733582] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356872, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.878162] env[61906]: DEBUG oslo_concurrency.lockutils [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.878461] env[61906]: DEBUG oslo_concurrency.lockutils [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.878723] env[61906]: DEBUG nova.objects.instance [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lazy-loading 'resources' on Instance uuid bc773559-fe14-485f-8bdc-1d4aab186521 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.130929] env[61906]: DEBUG oslo_concurrency.lockutils [None req-044c23ba-03ef-433f-bbc2-34cf4d9029ba tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "e35580b3-7177-46fd-bb86-b013efbf4911" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.465s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.208739] env[61906]: DEBUG oslo_vmware.api [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356870, 'name': ReconfigVM_Task, 'duration_secs': 0.776076} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.209098] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Reconfigured VM instance instance-00000051 to attach disk [datastore2] volume-cfc83076-1277-4f42-9c90-ef2e9aec2b8a/volume-cfc83076-1277-4f42-9c90-ef2e9aec2b8a.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.213743] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c903cb4-5d6d-4dfa-9045-5baaf4500a7b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.230086] env[61906]: DEBUG oslo_vmware.api [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 913.230086] env[61906]: value = "task-1356873" [ 913.230086] env[61906]: _type = "Task" [ 913.230086] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.235935] env[61906]: DEBUG oslo_vmware.api [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356872, 'name': PowerOnVM_Task, 'duration_secs': 0.415527} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.236564] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 913.236787] env[61906]: INFO nova.compute.manager [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Took 7.63 seconds to spawn the instance on the hypervisor. [ 913.236968] env[61906]: DEBUG nova.compute.manager [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 913.237727] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2c4bad-1ed7-4e87-a367-1967294946df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.243155] env[61906]: DEBUG oslo_vmware.api [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356873, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.450028] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Updating instance_info_cache with network_info: [{"id": "e291dedf-5c3f-4d78-bad3-295b2d1e2f75", "address": "fa:16:3e:c8:8d:ef", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape291dedf-5c", "ovs_interfaceid": "e291dedf-5c3f-4d78-bad3-295b2d1e2f75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.521481] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c9f1de-c553-463d-831f-58d9560fb876 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.528840] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad91fbba-91c4-4b5a-b112-846529ac2418 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.557898] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7f95fa-1936-43b8-8962-9921aebb2b07 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.565565] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda8bcd4-e980-430d-8443-d1442ef96c81 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.578303] env[61906]: DEBUG nova.compute.provider_tree [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.592552] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "e35580b3-7177-46fd-bb86-b013efbf4911" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.592791] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "e35580b3-7177-46fd-bb86-b013efbf4911" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.593014] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "e35580b3-7177-46fd-bb86-b013efbf4911-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.594385] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "e35580b3-7177-46fd-bb86-b013efbf4911-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.594385] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "e35580b3-7177-46fd-bb86-b013efbf4911-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.595291] env[61906]: INFO nova.compute.manager [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Terminating instance [ 913.598383] env[61906]: DEBUG nova.compute.manager [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 913.598383] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 913.598383] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eca52c0-27a4-4cad-9403-d8a27390cbc9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.605350] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.605594] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-535a4510-ff12-4d16-9ace-aab421abbcc7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.613270] env[61906]: DEBUG oslo_vmware.api [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 913.613270] env[61906]: value = "task-1356874" [ 913.613270] env[61906]: _type = "Task" [ 913.613270] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.619074] env[61906]: DEBUG oslo_vmware.api [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356874, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.669476] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.739872] env[61906]: DEBUG oslo_vmware.api [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356873, 'name': ReconfigVM_Task, 'duration_secs': 0.172376} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.741025] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-289013', 'volume_id': 'cfc83076-1277-4f42-9c90-ef2e9aec2b8a', 'name': 'volume-cfc83076-1277-4f42-9c90-ef2e9aec2b8a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63', 'attached_at': '', 'detached_at': '', 'volume_id': 'cfc83076-1277-4f42-9c90-ef2e9aec2b8a', 'serial': 'cfc83076-1277-4f42-9c90-ef2e9aec2b8a'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 913.759464] env[61906]: INFO nova.compute.manager [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Took 14.15 seconds to build instance. [ 913.952812] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Releasing lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.953043] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 913.953247] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.953500] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.953689] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.953877] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.954089] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.954245] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.954374] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 913.954520] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 914.082694] env[61906]: DEBUG nova.scheduler.client.report [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.121784] env[61906]: DEBUG oslo_vmware.api [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356874, 'name': PowerOffVM_Task, 'duration_secs': 0.18433} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.122078] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 914.122257] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 914.122508] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-577b81e6-6719-4cc2-b512-8eaa6af29226 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.196046] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.196046] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.196046] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleting the datastore file [datastore1] e35580b3-7177-46fd-bb86-b013efbf4911 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.196280] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e6e7eae0-70c2-44e9-b79a-6d3d0fadac5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.202835] env[61906]: DEBUG oslo_vmware.api [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 914.202835] env[61906]: value = "task-1356876" [ 914.202835] env[61906]: _type = "Task" [ 914.202835] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.211040] env[61906]: DEBUG oslo_vmware.api [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356876, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.260394] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41a74236-2f36-418a-9716-1b12ddda209a tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.662s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.357061] env[61906]: DEBUG oslo_concurrency.lockutils [None req-32840d8f-986a-42b6-9e51-27e2b0f409da tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.357061] env[61906]: DEBUG oslo_concurrency.lockutils [None req-32840d8f-986a-42b6-9e51-27e2b0f409da tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.357315] env[61906]: DEBUG nova.compute.manager [None req-32840d8f-986a-42b6-9e51-27e2b0f409da tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.358273] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004da200-b9db-40cb-ba10-3bf5d17f35d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.366569] env[61906]: DEBUG nova.compute.manager [None req-32840d8f-986a-42b6-9e51-27e2b0f409da tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61906) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 914.367229] env[61906]: DEBUG nova.objects.instance [None req-32840d8f-986a-42b6-9e51-27e2b0f409da tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lazy-loading 'flavor' on Instance uuid a4c9f463-98e7-49eb-8c01-c8f0ee98a884 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.458163] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.588346] env[61906]: DEBUG oslo_concurrency.lockutils [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.710s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.590728] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.921s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.590971] env[61906]: DEBUG nova.objects.instance [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lazy-loading 'pci_requests' on Instance uuid cb879dc0-3af7-4279-aa28-66e2b8b4286d {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.610229] env[61906]: INFO nova.scheduler.client.report [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Deleted allocations for instance bc773559-fe14-485f-8bdc-1d4aab186521 [ 914.714073] env[61906]: DEBUG oslo_vmware.api [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356876, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.272746} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.714442] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.714697] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 914.714959] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 914.715255] env[61906]: INFO nova.compute.manager [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Took 1.12 seconds to destroy the instance on the hypervisor. [ 914.715610] env[61906]: DEBUG oslo.service.loopingcall [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.715885] env[61906]: DEBUG nova.compute.manager [-] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 914.716028] env[61906]: DEBUG nova.network.neutron [-] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 914.730378] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "e133605d-c630-4b9e-a314-bf496c853710" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.730704] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.731029] env[61906]: INFO nova.compute.manager [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Rebooting instance [ 914.777497] env[61906]: DEBUG nova.objects.instance [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lazy-loading 'flavor' on Instance uuid 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.873624] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-32840d8f-986a-42b6-9e51-27e2b0f409da tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.874604] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2cfe425d-53cf-4e04-8817-5ee2e3c20f8e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.880844] env[61906]: DEBUG oslo_vmware.api [None req-32840d8f-986a-42b6-9e51-27e2b0f409da tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 914.880844] env[61906]: value = "task-1356877" [ 914.880844] env[61906]: _type = "Task" [ 914.880844] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.888912] env[61906]: DEBUG oslo_vmware.api [None req-32840d8f-986a-42b6-9e51-27e2b0f409da tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356877, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.946285] env[61906]: DEBUG nova.compute.manager [req-9b7e6db2-e6fb-45b6-a678-adef9535dfc7 req-d44c62de-82a3-439f-803d-2bc94b9425b4 service nova] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Received event network-vif-deleted-2f799c25-3b8b-4759-aadf-35fc1e98022e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.946658] env[61906]: INFO nova.compute.manager [req-9b7e6db2-e6fb-45b6-a678-adef9535dfc7 req-d44c62de-82a3-439f-803d-2bc94b9425b4 service nova] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Neutron deleted interface 2f799c25-3b8b-4759-aadf-35fc1e98022e; detaching it from the instance and deleting it from the info cache [ 914.948022] env[61906]: DEBUG nova.network.neutron [req-9b7e6db2-e6fb-45b6-a678-adef9535dfc7 req-d44c62de-82a3-439f-803d-2bc94b9425b4 service nova] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.098506] env[61906]: DEBUG nova.objects.instance [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lazy-loading 'numa_topology' on Instance uuid cb879dc0-3af7-4279-aa28-66e2b8b4286d {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.117804] env[61906]: DEBUG oslo_concurrency.lockutils [None req-09bd9784-48d3-42cc-809f-75bef236cdcd tempest-ServerDiskConfigTestJSON-854431548 tempest-ServerDiskConfigTestJSON-854431548-project-member] Lock "bc773559-fe14-485f-8bdc-1d4aab186521" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.343s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.254183] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.254428] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.254634] env[61906]: DEBUG nova.network.neutron [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 915.282603] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fe23aed4-2546-4401-8c51-3b5b7dc8c6d1 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.764s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.390763] env[61906]: DEBUG oslo_vmware.api [None req-32840d8f-986a-42b6-9e51-27e2b0f409da tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356877, 'name': PowerOffVM_Task, 'duration_secs': 0.184567} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.391401] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-32840d8f-986a-42b6-9e51-27e2b0f409da tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 915.395044] env[61906]: DEBUG nova.compute.manager [None req-32840d8f-986a-42b6-9e51-27e2b0f409da tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.395044] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45f1e55-9049-4469-b2bb-5aa4c2d5f2cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.423507] env[61906]: DEBUG nova.network.neutron [-] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.450125] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c97a6585-4089-4845-9d4b-c1fd89996b10 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.463898] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14401a90-3724-413d-a123-6f83ad9d05ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.496131] env[61906]: DEBUG nova.compute.manager [req-9b7e6db2-e6fb-45b6-a678-adef9535dfc7 req-d44c62de-82a3-439f-803d-2bc94b9425b4 service nova] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Detach interface failed, port_id=2f799c25-3b8b-4759-aadf-35fc1e98022e, reason: Instance e35580b3-7177-46fd-bb86-b013efbf4911 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 915.606123] env[61906]: INFO nova.compute.claims [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.907536] env[61906]: INFO nova.compute.manager [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Rescuing [ 915.907922] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.907922] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.908332] env[61906]: DEBUG nova.network.neutron [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 915.909622] env[61906]: DEBUG oslo_concurrency.lockutils [None req-32840d8f-986a-42b6-9e51-27e2b0f409da tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.553s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.924682] env[61906]: INFO nova.compute.manager [-] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Took 1.21 seconds to deallocate network for instance. [ 916.313613] env[61906]: DEBUG nova.network.neutron [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance_info_cache with network_info: [{"id": "d803c8e8-7145-412c-aa3e-2545666742eb", "address": "fa:16:3e:46:59:77", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd803c8e8-71", "ovs_interfaceid": "d803c8e8-7145-412c-aa3e-2545666742eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.431525] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.764043] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74135458-3e5a-4318-a0ab-6e3aebed2dbd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.774159] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5afefa7-2348-4efa-a263-1d3aa458d3dd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.804295] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65c83ba-f9a8-45fb-9b6b-734b2390dcaa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.811506] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6168489-db60-446c-ab73-33a7bd18f12c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.815453] env[61906]: DEBUG oslo_concurrency.lockutils [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.815672] env[61906]: DEBUG oslo_concurrency.lockutils [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.815877] env[61906]: DEBUG oslo_concurrency.lockutils [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.816079] env[61906]: DEBUG oslo_concurrency.lockutils [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.816256] env[61906]: DEBUG oslo_concurrency.lockutils [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.818215] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.822280] env[61906]: INFO nova.compute.manager [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Terminating instance [ 916.824373] env[61906]: DEBUG nova.compute.manager [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 916.824761] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.825504] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9656cd0c-a693-4433-8927-7ed4c40a3bb6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.841203] env[61906]: DEBUG nova.compute.provider_tree [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.847407] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.848096] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7ec19062-7911-44f3-8021-af93292bd282 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.850497] env[61906]: DEBUG nova.network.neutron [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updating instance_info_cache with network_info: [{"id": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "address": "fa:16:3e:f0:1b:68", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb78ba42-03", "ovs_interfaceid": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.854543] env[61906]: DEBUG nova.compute.manager [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.855327] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d7eb7c-7051-4d24-b013-f368a74b471c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.909673] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 916.909673] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 916.910101] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleting the datastore file [datastore2] a4c9f463-98e7-49eb-8c01-c8f0ee98a884 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.910433] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-235351d1-e5d8-499b-a994-e6277bcb7de8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.918272] env[61906]: DEBUG oslo_vmware.api [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 916.918272] env[61906]: value = "task-1356879" [ 916.918272] env[61906]: _type = "Task" [ 916.918272] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.927769] env[61906]: DEBUG oslo_vmware.api [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356879, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.348660] env[61906]: DEBUG nova.scheduler.client.report [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.354793] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.428971] env[61906]: DEBUG oslo_vmware.api [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356879, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139891} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.429234] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.429422] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.429606] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.429783] env[61906]: INFO nova.compute.manager [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Took 0.61 seconds to destroy the instance on the hypervisor. [ 917.430036] env[61906]: DEBUG oslo.service.loopingcall [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.430468] env[61906]: DEBUG nova.compute.manager [-] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 917.430468] env[61906]: DEBUG nova.network.neutron [-] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 917.710011] env[61906]: DEBUG nova.compute.manager [req-dd610f97-8b5a-478c-864f-9350f16649be req-4c83f0d3-a241-476c-88ab-a2c5e0b073ee service nova] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Received event network-vif-deleted-505e72fa-25c4-4cc1-981f-d34507bb4b97 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.710224] env[61906]: INFO nova.compute.manager [req-dd610f97-8b5a-478c-864f-9350f16649be req-4c83f0d3-a241-476c-88ab-a2c5e0b073ee service nova] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Neutron deleted interface 505e72fa-25c4-4cc1-981f-d34507bb4b97; detaching it from the instance and deleting it from the info cache [ 917.710391] env[61906]: DEBUG nova.network.neutron [req-dd610f97-8b5a-478c-864f-9350f16649be req-4c83f0d3-a241-476c-88ab-a2c5e0b073ee service nova] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.854384] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.264s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.856493] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.399s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.856673] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.856826] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 917.857269] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.426s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.857506] env[61906]: DEBUG nova.objects.instance [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lazy-loading 'resources' on Instance uuid e35580b3-7177-46fd-bb86-b013efbf4911 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.859686] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60ae6ac-cb60-492e-9c6a-f7f73f325249 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.868250] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ef08c5-3d10-4901-9e58-85cae2d2a88b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.873502] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2caf79-1b83-4d1d-9c0b-6aa5d4367499 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.890517] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.891323] env[61906]: INFO nova.network.neutron [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updating port f119e802-0a46-4390-927d-805c2e23a4b4 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 917.893888] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6ffb0e-444b-4a66-9de8-c85b476da072 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.896425] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Doing hard reboot of VM {{(pid=61906) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 917.896632] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67ec1658-e570-419c-b1de-c0614658263a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.899471] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-155ddaea-596e-43eb-900c-c35793f54453 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.904961] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6fad292-5874-4fb4-bf5f-159cfad66b2e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.909571] env[61906]: DEBUG oslo_vmware.api [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 917.909571] env[61906]: value = "task-1356881" [ 917.909571] env[61906]: _type = "Task" [ 917.909571] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.910030] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 917.910030] env[61906]: value = "task-1356880" [ 917.910030] env[61906]: _type = "Task" [ 917.910030] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.944041] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180419MB free_disk=150GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 917.944187] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.949761] env[61906]: DEBUG oslo_vmware.api [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356881, 'name': ResetVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.953103] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356880, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.190273] env[61906]: DEBUG nova.network.neutron [-] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.212981] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c647b292-bc01-4026-8ad4-4722867b11b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.222999] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36479542-f901-4029-b7b3-15b0e7213c67 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.249437] env[61906]: DEBUG nova.compute.manager [req-dd610f97-8b5a-478c-864f-9350f16649be req-4c83f0d3-a241-476c-88ab-a2c5e0b073ee service nova] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Detach interface failed, port_id=505e72fa-25c4-4cc1-981f-d34507bb4b97, reason: Instance a4c9f463-98e7-49eb-8c01-c8f0ee98a884 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 918.427128] env[61906]: DEBUG oslo_vmware.api [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356881, 'name': ResetVM_Task, 'duration_secs': 0.103857} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.430171] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Did hard reboot of VM {{(pid=61906) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 918.430399] env[61906]: DEBUG nova.compute.manager [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.430693] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356880, 'name': PowerOffVM_Task, 'duration_secs': 0.21497} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.431705] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb61b815-6435-4126-9b63-1836c1fc5a96 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.433725] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.438164] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54736f4-8aaa-4b22-bd11-ce0f38b6d586 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.465334] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8de1e7-48bd-4ec1-b028-4fec4de57b80 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.500183] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 918.500750] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cad219ac-d785-4d95-a3d6-5eb7d0fc7c88 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.509565] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 918.509565] env[61906]: value = "task-1356882" [ 918.509565] env[61906]: _type = "Task" [ 918.509565] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.517406] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356882, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.529397] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b4e684-f88f-44c5-9895-9cc2c8346e53 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.536148] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8782bccd-82a1-4b54-8e53-fe832fa084a3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.569021] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c79eae-205c-4387-9a6e-8568c349da92 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.575681] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef3549d-7871-4484-9833-4f990e0ff832 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.590431] env[61906]: DEBUG nova.compute.provider_tree [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.694668] env[61906]: INFO nova.compute.manager [-] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Took 1.26 seconds to deallocate network for instance. [ 918.952225] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9314f6c7-831f-4b85-9527-3718dfccc6b4 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.221s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.019673] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] VM already powered off {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 919.019894] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.020157] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.020313] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.020500] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.020760] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0fcfaebb-92e0-4191-8f11-e842b2a70038 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.029727] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.029940] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 919.030648] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a5c08ce-5f37-44fe-b866-2569df457375 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.039159] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 919.039159] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5234c858-450f-09b9-e4de-8fc1e83fe285" [ 919.039159] env[61906]: _type = "Task" [ 919.039159] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.047637] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5234c858-450f-09b9-e4de-8fc1e83fe285, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.093823] env[61906]: DEBUG nova.scheduler.client.report [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.201674] env[61906]: DEBUG oslo_concurrency.lockutils [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.549439] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5234c858-450f-09b9-e4de-8fc1e83fe285, 'name': SearchDatastore_Task, 'duration_secs': 0.009517} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.550176] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac34902d-f95d-45a2-8c2a-13bf77903aa6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.556076] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 919.556076] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525666bb-9328-8314-4ece-c1d848396630" [ 919.556076] env[61906]: _type = "Task" [ 919.556076] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.564305] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525666bb-9328-8314-4ece-c1d848396630, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.598445] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.741s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.602119] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.658s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.612786] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.612893] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.613140] env[61906]: DEBUG nova.network.neutron [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 919.628403] env[61906]: INFO nova.scheduler.client.report [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleted allocations for instance e35580b3-7177-46fd-bb86-b013efbf4911 [ 919.741231] env[61906]: DEBUG nova.compute.manager [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Received event network-vif-plugged-f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.741474] env[61906]: DEBUG oslo_concurrency.lockutils [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] Acquiring lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.741685] env[61906]: DEBUG oslo_concurrency.lockutils [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.741859] env[61906]: DEBUG oslo_concurrency.lockutils [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.742582] env[61906]: DEBUG nova.compute.manager [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] No waiting events found dispatching network-vif-plugged-f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 919.742582] env[61906]: WARNING nova.compute.manager [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Received unexpected event network-vif-plugged-f119e802-0a46-4390-927d-805c2e23a4b4 for instance with vm_state shelved_offloaded and task_state spawning. [ 919.742582] env[61906]: DEBUG nova.compute.manager [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Received event network-changed-f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.742582] env[61906]: DEBUG nova.compute.manager [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Refreshing instance network info cache due to event network-changed-f119e802-0a46-4390-927d-805c2e23a4b4. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 919.743134] env[61906]: DEBUG oslo_concurrency.lockutils [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] Acquiring lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.069059] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525666bb-9328-8314-4ece-c1d848396630, 'name': SearchDatastore_Task, 'duration_secs': 0.009109} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.069355] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.069699] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1-rescue.vmdk. {{(pid=61906) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 920.069916] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec3063dd-499b-40fe-881f-619deba0300c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.076530] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 920.076530] env[61906]: value = "task-1356883" [ 920.076530] env[61906]: _type = "Task" [ 920.076530] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.086980] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356883, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.139717] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e7a0a0e6-6f7e-404c-a575-537215a0eb42 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "e35580b3-7177-46fd-bb86-b013efbf4911" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.547s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.410817] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "bb53f018-8546-4a00-a4be-d3c48416b1a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.411878] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "bb53f018-8546-4a00-a4be-d3c48416b1a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.411878] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "bb53f018-8546-4a00-a4be-d3c48416b1a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.411878] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "bb53f018-8546-4a00-a4be-d3c48416b1a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.411878] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "bb53f018-8546-4a00-a4be-d3c48416b1a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.418161] env[61906]: INFO nova.compute.manager [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Terminating instance [ 920.422087] env[61906]: DEBUG nova.compute.manager [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 920.422395] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 920.423498] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-987d9a4a-6edb-488e-89d5-b2679c848ce3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.437746] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.439139] env[61906]: DEBUG nova.network.neutron [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updating instance_info_cache with network_info: [{"id": "f119e802-0a46-4390-927d-805c2e23a4b4", "address": "fa:16:3e:2c:b5:5c", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119e802-0a", "ovs_interfaceid": "f119e802-0a46-4390-927d-805c2e23a4b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.441231] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87963be0-3f73-445f-b3a8-54a6c61a2891 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.451227] env[61906]: DEBUG oslo_vmware.api [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 920.451227] env[61906]: value = "task-1356884" [ 920.451227] env[61906]: _type = "Task" [ 920.451227] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.465809] env[61906]: DEBUG oslo_vmware.api [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356884, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.588797] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356883, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485262} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.589820] env[61906]: INFO nova.virt.vmwareapi.ds_util [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1-rescue.vmdk. [ 920.589959] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46e5d0d-573b-4602-9018-6e0322e7ded7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.623771] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.626373] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f25ce26f-085c-4993-8330-1c815ed66d4f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.645057] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 920.645057] env[61906]: value = "task-1356885" [ 920.645057] env[61906]: _type = "Task" [ 920.645057] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.653789] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356885, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.670602] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 79452791-59cb-4722-bb4a-8e59d8c4e641 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.670793] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 7bae658e-b050-4639-b34e-c2671ef5c773 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.670996] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance e133605d-c630-4b9e-a314-bf496c853710 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.671210] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.671210] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance bb53f018-8546-4a00-a4be-d3c48416b1a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.672051] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 99e5c55d-79ae-47a6-8500-79fc68291650 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.672051] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance d060a9f7-bd96-4e95-8780-1617a6ca7443 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.672051] env[61906]: WARNING nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance a4c9f463-98e7-49eb-8c01-c8f0ee98a884 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 920.672051] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance cb879dc0-3af7-4279-aa28-66e2b8b4286d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.947307] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.949184] env[61906]: DEBUG oslo_concurrency.lockutils [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] Acquired lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.949184] env[61906]: DEBUG nova.network.neutron [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Refreshing network info cache for port f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 920.969060] env[61906]: DEBUG oslo_vmware.api [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356884, 'name': PowerOffVM_Task, 'duration_secs': 0.406224} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.969060] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 920.969060] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 920.969256] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-663c0bc8-5127-486b-98f5-4770502d4ba2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.983222] env[61906]: DEBUG nova.virt.hardware [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='f29fd203e627becc9deaa7e59bfc6f92',container_format='bare',created_at=2024-10-24T14:11:09Z,direct_url=,disk_format='vmdk',id=4e467722-4f47-4b26-9f34-13132817ea0e,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-127540343-shelved',owner='8f3bb9ce83b3430bb9202d3c72e77b4c',properties=ImageMetaProps,protected=,size=31669760,status='active',tags=,updated_at=2024-10-24T14:11:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 920.983515] env[61906]: DEBUG nova.virt.hardware [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 920.983732] env[61906]: DEBUG nova.virt.hardware [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 920.983940] env[61906]: DEBUG nova.virt.hardware [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 920.984112] env[61906]: DEBUG nova.virt.hardware [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 920.984285] env[61906]: DEBUG nova.virt.hardware [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 920.984485] env[61906]: DEBUG nova.virt.hardware [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 920.984657] env[61906]: DEBUG nova.virt.hardware [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 920.984837] env[61906]: DEBUG nova.virt.hardware [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 920.985015] env[61906]: DEBUG nova.virt.hardware [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 920.985242] env[61906]: DEBUG nova.virt.hardware [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.986563] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833c1b2b-0b62-4c3b-a8cc-f581d2ce3afa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.998462] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe87bd4-114f-480e-a8b3-f261ac6e0530 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.015078] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:b5:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8edfde4-5a99-4745-956d-04da82ab1b85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f119e802-0a46-4390-927d-805c2e23a4b4', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.023261] env[61906]: DEBUG oslo.service.loopingcall [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.023667] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 921.024261] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-082c7af3-efd0-4915-8f1a-df6cbffc50e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.052203] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 921.052203] env[61906]: value = "task-1356887" [ 921.052203] env[61906]: _type = "Task" [ 921.052203] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.055167] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.055430] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.055620] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleting the datastore file [datastore2] bb53f018-8546-4a00-a4be-d3c48416b1a1 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.057592] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d4ae324c-68d7-4733-b475-711c0347439a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.062113] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356887, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.068668] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "070f8882-9099-4f17-bae8-1701b5dcc08d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.068903] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "070f8882-9099-4f17-bae8-1701b5dcc08d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.070655] env[61906]: DEBUG oslo_vmware.api [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 921.070655] env[61906]: value = "task-1356888" [ 921.070655] env[61906]: _type = "Task" [ 921.070655] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.079345] env[61906]: DEBUG oslo_vmware.api [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356888, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.158410] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356885, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.179030] env[61906]: INFO nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 070f8882-9099-4f17-bae8-1701b5dcc08d has allocations against this compute host but is not found in the database. [ 921.179030] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 921.179030] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 921.368404] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884015e1-df84-49e2-a598-1fe34ba3e4b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.375937] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f88396-2711-4ffc-a159-736ced209c9b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.410060] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76689a2-7dc3-4e78-a117-29d0f0022307 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.418019] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe85dfa-92bb-48ac-a7d3-769e7d971286 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.433654] env[61906]: DEBUG nova.compute.provider_tree [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.561095] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356887, 'name': CreateVM_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.576981] env[61906]: DEBUG nova.compute.manager [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 921.589861] env[61906]: DEBUG oslo_vmware.api [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356888, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.335599} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.590197] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 921.590429] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 921.590651] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 921.590867] env[61906]: INFO nova.compute.manager [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Took 1.17 seconds to destroy the instance on the hypervisor. [ 921.591171] env[61906]: DEBUG oslo.service.loopingcall [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.591412] env[61906]: DEBUG nova.compute.manager [-] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 921.591521] env[61906]: DEBUG nova.network.neutron [-] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 921.656578] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356885, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.939373] env[61906]: DEBUG nova.scheduler.client.report [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.027461] env[61906]: DEBUG nova.network.neutron [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updated VIF entry in instance network info cache for port f119e802-0a46-4390-927d-805c2e23a4b4. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 922.027832] env[61906]: DEBUG nova.network.neutron [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updating instance_info_cache with network_info: [{"id": "f119e802-0a46-4390-927d-805c2e23a4b4", "address": "fa:16:3e:2c:b5:5c", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf119e802-0a", "ovs_interfaceid": "f119e802-0a46-4390-927d-805c2e23a4b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.062160] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356887, 'name': CreateVM_Task, 'duration_secs': 0.652578} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.062345] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 922.063093] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4e467722-4f47-4b26-9f34-13132817ea0e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.063314] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4e467722-4f47-4b26-9f34-13132817ea0e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.063673] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/4e467722-4f47-4b26-9f34-13132817ea0e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 922.063941] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb0895d2-5876-4f7f-a454-af481f3d5fe8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.068779] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 922.068779] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a02740-01f0-9d18-f324-068b158e4f88" [ 922.068779] env[61906]: _type = "Task" [ 922.068779] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.078017] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a02740-01f0-9d18-f324-068b158e4f88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.116076] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.158740] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356885, 'name': ReconfigVM_Task, 'duration_secs': 1.047424} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.160126] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1-rescue.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.161539] env[61906]: DEBUG nova.compute.manager [req-2df8bdf2-0a04-4b72-8857-574276459756 req-4d8fdd24-e6ce-402b-afaf-662c69fe7225 service nova] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Received event network-vif-deleted-7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.161819] env[61906]: INFO nova.compute.manager [req-2df8bdf2-0a04-4b72-8857-574276459756 req-4d8fdd24-e6ce-402b-afaf-662c69fe7225 service nova] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Neutron deleted interface 7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6; detaching it from the instance and deleting it from the info cache [ 922.162798] env[61906]: DEBUG nova.network.neutron [req-2df8bdf2-0a04-4b72-8857-574276459756 req-4d8fdd24-e6ce-402b-afaf-662c69fe7225 service nova] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.163586] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c956331-1c6a-4121-9f53-57dcd12ecd40 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.196512] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9799f814-0ab9-40df-ae38-9d9be274f107 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.215105] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 922.215105] env[61906]: value = "task-1356889" [ 922.215105] env[61906]: _type = "Task" [ 922.215105] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.225838] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356889, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.449450] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 922.449690] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.848s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.450065] env[61906]: DEBUG oslo_concurrency.lockutils [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.249s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.450203] env[61906]: DEBUG oslo_concurrency.lockutils [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.452558] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.337s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.454250] env[61906]: INFO nova.compute.claims [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 922.482504] env[61906]: INFO nova.scheduler.client.report [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleted allocations for instance a4c9f463-98e7-49eb-8c01-c8f0ee98a884 [ 922.530717] env[61906]: DEBUG oslo_concurrency.lockutils [req-1ee0a084-126a-4ead-a83b-0ce2b8ed14ef req-088bdb07-7d05-426f-9867-a4b20231d365 service nova] Releasing lock "refresh_cache-cb879dc0-3af7-4279-aa28-66e2b8b4286d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.579200] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4e467722-4f47-4b26-9f34-13132817ea0e" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.579555] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Processing image 4e467722-4f47-4b26-9f34-13132817ea0e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.579721] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/4e467722-4f47-4b26-9f34-13132817ea0e/4e467722-4f47-4b26-9f34-13132817ea0e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.579878] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "[datastore1] devstack-image-cache_base/4e467722-4f47-4b26-9f34-13132817ea0e/4e467722-4f47-4b26-9f34-13132817ea0e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.580083] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.580340] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-886e6812-49aa-407b-b865-eb7727eaa3d9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.588306] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.588724] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 922.589152] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ed149c1-412d-47a6-aee6-ccfac9ccedf2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.594235] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 922.594235] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529ea1d0-85a5-d4ab-0081-f6a7d27b0789" [ 922.594235] env[61906]: _type = "Task" [ 922.594235] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.601642] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529ea1d0-85a5-d4ab-0081-f6a7d27b0789, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.629150] env[61906]: DEBUG nova.network.neutron [-] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.668141] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4d84e5d8-e47d-44db-a321-6e4f63726d89 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.680967] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780a14f5-9d08-41da-8a96-f84afc4d9b28 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.707836] env[61906]: DEBUG nova.compute.manager [req-2df8bdf2-0a04-4b72-8857-574276459756 req-4d8fdd24-e6ce-402b-afaf-662c69fe7225 service nova] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Detach interface failed, port_id=7c4d0cd2-9fb7-40ee-9753-c735e62bf1c6, reason: Instance bb53f018-8546-4a00-a4be-d3c48416b1a1 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 922.723313] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356889, 'name': ReconfigVM_Task, 'duration_secs': 0.172435} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.723592] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 922.723895] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d3e3abe-7d1f-432b-aeb5-ab35836e7faa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.730433] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 922.730433] env[61906]: value = "task-1356890" [ 922.730433] env[61906]: _type = "Task" [ 922.730433] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.740939] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356890, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.991090] env[61906]: DEBUG oslo_concurrency.lockutils [None req-87342c64-7bf9-4c21-bada-e3f41cfd0fa3 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "a4c9f463-98e7-49eb-8c01-c8f0ee98a884" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.175s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.107792] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Preparing fetch location {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 923.108204] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Fetch image to [datastore1] OSTACK_IMG_24a73825-6eed-45e1-a549-789d1098db8e/OSTACK_IMG_24a73825-6eed-45e1-a549-789d1098db8e.vmdk {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 923.108486] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Downloading stream optimized image 4e467722-4f47-4b26-9f34-13132817ea0e to [datastore1] OSTACK_IMG_24a73825-6eed-45e1-a549-789d1098db8e/OSTACK_IMG_24a73825-6eed-45e1-a549-789d1098db8e.vmdk on the data store datastore1 as vApp {{(pid=61906) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 923.108753] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Downloading image file data 4e467722-4f47-4b26-9f34-13132817ea0e to the ESX as VM named 'OSTACK_IMG_24a73825-6eed-45e1-a549-789d1098db8e' {{(pid=61906) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 923.131211] env[61906]: INFO nova.compute.manager [-] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Took 1.54 seconds to deallocate network for instance. [ 923.190208] env[61906]: DEBUG oslo_vmware.rw_handles [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 923.190208] env[61906]: value = "resgroup-9" [ 923.190208] env[61906]: _type = "ResourcePool" [ 923.190208] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 923.190602] env[61906]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-dab4e4fc-b867-47ca-868b-33b1baa803f8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.213479] env[61906]: DEBUG oslo_vmware.rw_handles [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lease: (returnval){ [ 923.213479] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52cfa97f-96f6-cac0-0043-5ca12af859c3" [ 923.213479] env[61906]: _type = "HttpNfcLease" [ 923.213479] env[61906]: } obtained for vApp import into resource pool (val){ [ 923.213479] env[61906]: value = "resgroup-9" [ 923.213479] env[61906]: _type = "ResourcePool" [ 923.213479] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 923.213790] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the lease: (returnval){ [ 923.213790] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52cfa97f-96f6-cac0-0043-5ca12af859c3" [ 923.213790] env[61906]: _type = "HttpNfcLease" [ 923.213790] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 923.220253] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 923.220253] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52cfa97f-96f6-cac0-0043-5ca12af859c3" [ 923.220253] env[61906]: _type = "HttpNfcLease" [ 923.220253] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 923.238746] env[61906]: DEBUG oslo_vmware.api [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356890, 'name': PowerOnVM_Task, 'duration_secs': 0.395338} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.239035] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 923.241745] env[61906]: DEBUG nova.compute.manager [None req-74b9c48b-a363-4d25-9740-7ef9109ea754 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.242763] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f8cbdd-c5c4-48e0-a790-245fe4fedef0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.598229] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52add01e-6e45-4bb0-9b5d-8baf7a6740bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.606260] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e114af6-18b0-4701-9de0-b9d624cf1ddc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.637195] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17e93a8-1668-49d6-9508-ca3c287fd26f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.645053] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb83dd9-eab9-45fa-99b2-005f651a2d8b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.661067] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.661506] env[61906]: DEBUG nova.compute.provider_tree [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.670690] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquiring lock "c77823f1-189c-4ac9-9a40-7bc45e82b442" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.670919] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Lock "c77823f1-189c-4ac9-9a40-7bc45e82b442" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.722224] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 923.722224] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52cfa97f-96f6-cac0-0043-5ca12af859c3" [ 923.722224] env[61906]: _type = "HttpNfcLease" [ 923.722224] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 923.908876] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "766e79cb-bea3-4b74-923d-a147c7be4134" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.909256] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "766e79cb-bea3-4b74-923d-a147c7be4134" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.165244] env[61906]: DEBUG nova.scheduler.client.report [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.174541] env[61906]: DEBUG nova.compute.manager [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 924.222487] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 924.222487] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52cfa97f-96f6-cac0-0043-5ca12af859c3" [ 924.222487] env[61906]: _type = "HttpNfcLease" [ 924.222487] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 924.308710] env[61906]: INFO nova.compute.manager [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Unrescuing [ 924.309099] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.309350] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquired lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.309577] env[61906]: DEBUG nova.network.neutron [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 924.411496] env[61906]: DEBUG nova.compute.manager [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 924.669954] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.218s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.670457] env[61906]: DEBUG nova.compute.manager [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 924.674709] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.014s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.675665] env[61906]: DEBUG nova.objects.instance [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lazy-loading 'resources' on Instance uuid bb53f018-8546-4a00-a4be-d3c48416b1a1 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.697368] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.723038] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 924.723038] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52cfa97f-96f6-cac0-0043-5ca12af859c3" [ 924.723038] env[61906]: _type = "HttpNfcLease" [ 924.723038] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 924.723335] env[61906]: DEBUG oslo_vmware.rw_handles [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 924.723335] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52cfa97f-96f6-cac0-0043-5ca12af859c3" [ 924.723335] env[61906]: _type = "HttpNfcLease" [ 924.723335] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 924.724029] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4147ec0-bd00-48ba-94de-d39762b6b597 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.730996] env[61906]: DEBUG oslo_vmware.rw_handles [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e551e-d815-7427-a620-7315410de5f4/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 924.731191] env[61906]: DEBUG oslo_vmware.rw_handles [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating HTTP connection to write to file with size = 31669760 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e551e-d815-7427-a620-7315410de5f4/disk-0.vmdk. {{(pid=61906) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 924.795234] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-66cbb6a6-c739-4314-b89b-1158fa7623fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.931291] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.060695] env[61906]: DEBUG nova.network.neutron [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updating instance_info_cache with network_info: [{"id": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "address": "fa:16:3e:f0:1b:68", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb78ba42-03", "ovs_interfaceid": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.179047] env[61906]: DEBUG nova.compute.utils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 925.183531] env[61906]: DEBUG nova.compute.manager [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 925.183618] env[61906]: DEBUG nova.network.neutron [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 925.257381] env[61906]: DEBUG nova.policy [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ec22db431334e7886acb497ebfc7eb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ac9a6686e7184fb1913e84a77985b449', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 925.351096] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5dd4e98-9635-47a1-b7de-03820f76365d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.362567] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afbd879-57a2-4c3e-93e8-ee87051c0960 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.399275] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894c065e-273b-4d61-924f-049c36add252 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.409085] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68897919-0453-46a4-80de-16ae9905cdb6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.424891] env[61906]: DEBUG nova.compute.provider_tree [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.538638] env[61906]: DEBUG nova.network.neutron [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Successfully created port: e650d2e9-acd3-479b-9d48-748636ace077 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.563960] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Releasing lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.564757] env[61906]: DEBUG nova.objects.instance [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lazy-loading 'flavor' on Instance uuid 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.685802] env[61906]: DEBUG nova.compute.manager [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 925.921157] env[61906]: DEBUG oslo_vmware.rw_handles [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Completed reading data from the image iterator. {{(pid=61906) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 925.921386] env[61906]: DEBUG oslo_vmware.rw_handles [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e551e-d815-7427-a620-7315410de5f4/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 925.922366] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17add09c-8188-4b1c-80da-e9634da54553 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.928949] env[61906]: DEBUG oslo_vmware.rw_handles [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e551e-d815-7427-a620-7315410de5f4/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 925.929141] env[61906]: DEBUG oslo_vmware.rw_handles [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e551e-d815-7427-a620-7315410de5f4/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 925.930031] env[61906]: DEBUG nova.scheduler.client.report [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.933164] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-dbd2f3a1-bf89-40bc-8327-053ab552e722 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.070741] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb0e82c-7e3a-4c72-bed3-d9615c86e92c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.096086] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 926.096464] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9c63d8e-196d-4966-9921-0f75092f85f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.103872] env[61906]: DEBUG oslo_vmware.api [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 926.103872] env[61906]: value = "task-1356892" [ 926.103872] env[61906]: _type = "Task" [ 926.103872] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.111626] env[61906]: DEBUG oslo_vmware.api [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356892, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.305902] env[61906]: DEBUG oslo_vmware.rw_handles [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523e551e-d815-7427-a620-7315410de5f4/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 926.306300] env[61906]: INFO nova.virt.vmwareapi.images [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Downloaded image file data 4e467722-4f47-4b26-9f34-13132817ea0e [ 926.307450] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6259e28b-0b76-430b-a67e-d8c8dcfb9dbc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.323334] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff2aadd4-786d-49b8-9130-4b47a8ff7076 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.345429] env[61906]: INFO nova.virt.vmwareapi.images [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] The imported VM was unregistered [ 926.347868] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Caching image {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 926.348162] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating directory with path [datastore1] devstack-image-cache_base/4e467722-4f47-4b26-9f34-13132817ea0e {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.348464] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-59a31c78-ad34-4c03-8c04-683f6cf67ad7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.380783] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Created directory with path [datastore1] devstack-image-cache_base/4e467722-4f47-4b26-9f34-13132817ea0e {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.382030] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_24a73825-6eed-45e1-a549-789d1098db8e/OSTACK_IMG_24a73825-6eed-45e1-a549-789d1098db8e.vmdk to [datastore1] devstack-image-cache_base/4e467722-4f47-4b26-9f34-13132817ea0e/4e467722-4f47-4b26-9f34-13132817ea0e.vmdk. {{(pid=61906) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 926.382030] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-cd4294e9-e738-464d-b03f-0fa848ace6af {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.388288] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 926.388288] env[61906]: value = "task-1356894" [ 926.388288] env[61906]: _type = "Task" [ 926.388288] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.395694] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356894, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.436722] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.762s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.439121] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.742s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.440608] env[61906]: INFO nova.compute.claims [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 926.458147] env[61906]: INFO nova.scheduler.client.report [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleted allocations for instance bb53f018-8546-4a00-a4be-d3c48416b1a1 [ 926.613201] env[61906]: DEBUG oslo_vmware.api [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356892, 'name': PowerOffVM_Task, 'duration_secs': 0.24269} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.613531] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 926.618937] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Reconfiguring VM instance instance-00000051 to detach disk 2002 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 926.619235] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccdecb57-e075-4593-b3bc-71ad32031795 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.637113] env[61906]: DEBUG oslo_vmware.api [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 926.637113] env[61906]: value = "task-1356895" [ 926.637113] env[61906]: _type = "Task" [ 926.637113] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.644623] env[61906]: DEBUG oslo_vmware.api [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356895, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.694157] env[61906]: DEBUG nova.compute.manager [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 926.720264] env[61906]: DEBUG nova.virt.hardware [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.720577] env[61906]: DEBUG nova.virt.hardware [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.720839] env[61906]: DEBUG nova.virt.hardware [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.721365] env[61906]: DEBUG nova.virt.hardware [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.721365] env[61906]: DEBUG nova.virt.hardware [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.721501] env[61906]: DEBUG nova.virt.hardware [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.721634] env[61906]: DEBUG nova.virt.hardware [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.721840] env[61906]: DEBUG nova.virt.hardware [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.722042] env[61906]: DEBUG nova.virt.hardware [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.722228] env[61906]: DEBUG nova.virt.hardware [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.722498] env[61906]: DEBUG nova.virt.hardware [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.723499] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b7928a5-3900-4b36-87e4-59b6877ed101 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.732994] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7becbef-d8c4-4ba1-a1e7-3f8daeb4ab60 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.899266] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356894, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.965608] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73e61d87-1418-46d7-8a10-d4016232886c tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "bb53f018-8546-4a00-a4be-d3c48416b1a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.554s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.000759] env[61906]: DEBUG nova.compute.manager [req-54e884e0-de68-4c0b-8114-5f13e2f632ea req-5ce7cc96-d304-418f-816a-ddf23d6e005f service nova] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Received event network-vif-plugged-e650d2e9-acd3-479b-9d48-748636ace077 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.001046] env[61906]: DEBUG oslo_concurrency.lockutils [req-54e884e0-de68-4c0b-8114-5f13e2f632ea req-5ce7cc96-d304-418f-816a-ddf23d6e005f service nova] Acquiring lock "070f8882-9099-4f17-bae8-1701b5dcc08d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.001275] env[61906]: DEBUG oslo_concurrency.lockutils [req-54e884e0-de68-4c0b-8114-5f13e2f632ea req-5ce7cc96-d304-418f-816a-ddf23d6e005f service nova] Lock "070f8882-9099-4f17-bae8-1701b5dcc08d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.001511] env[61906]: DEBUG oslo_concurrency.lockutils [req-54e884e0-de68-4c0b-8114-5f13e2f632ea req-5ce7cc96-d304-418f-816a-ddf23d6e005f service nova] Lock "070f8882-9099-4f17-bae8-1701b5dcc08d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.001719] env[61906]: DEBUG nova.compute.manager [req-54e884e0-de68-4c0b-8114-5f13e2f632ea req-5ce7cc96-d304-418f-816a-ddf23d6e005f service nova] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] No waiting events found dispatching network-vif-plugged-e650d2e9-acd3-479b-9d48-748636ace077 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 927.001897] env[61906]: WARNING nova.compute.manager [req-54e884e0-de68-4c0b-8114-5f13e2f632ea req-5ce7cc96-d304-418f-816a-ddf23d6e005f service nova] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Received unexpected event network-vif-plugged-e650d2e9-acd3-479b-9d48-748636ace077 for instance with vm_state building and task_state spawning. [ 927.098953] env[61906]: DEBUG nova.network.neutron [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Successfully updated port: e650d2e9-acd3-479b-9d48-748636ace077 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 927.147681] env[61906]: DEBUG oslo_vmware.api [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356895, 'name': ReconfigVM_Task, 'duration_secs': 0.248943} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.148926] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Reconfigured VM instance instance-00000051 to detach disk 2002 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 927.148926] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.148926] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-20fe22b5-4446-49f9-a557-19d19c99f917 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.155887] env[61906]: DEBUG oslo_vmware.api [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 927.155887] env[61906]: value = "task-1356896" [ 927.155887] env[61906]: _type = "Task" [ 927.155887] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.164373] env[61906]: DEBUG oslo_vmware.api [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356896, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.400710] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356894, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.596340] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3462377f-45fc-409e-8e0d-fb80de071f62 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.603029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "refresh_cache-070f8882-9099-4f17-bae8-1701b5dcc08d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.603029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "refresh_cache-070f8882-9099-4f17-bae8-1701b5dcc08d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.603029] env[61906]: DEBUG nova.network.neutron [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 927.608679] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413ad7ef-a779-4f03-a68e-df113441d2db {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.641917] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c002d91e-1615-4634-8460-b23160da722a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.650309] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46787854-2a88-4918-bf21-8023a80a39a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.670242] env[61906]: DEBUG nova.compute.provider_tree [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.674794] env[61906]: DEBUG oslo_vmware.api [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356896, 'name': PowerOnVM_Task, 'duration_secs': 0.444776} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.675292] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 927.675576] env[61906]: DEBUG nova.compute.manager [None req-f3f14fa0-5ae8-484e-9284-b9c6eabfc098 tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.676337] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aad7331-340b-46fb-9fbf-de39c247a10b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.902503] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356894, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.140996] env[61906]: DEBUG nova.network.neutron [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 928.177155] env[61906]: DEBUG nova.scheduler.client.report [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.286483] env[61906]: DEBUG nova.network.neutron [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Updating instance_info_cache with network_info: [{"id": "e650d2e9-acd3-479b-9d48-748636ace077", "address": "fa:16:3e:71:a7:3c", "network": {"id": "5b29c9e0-23c9-42ee-9be4-60178bddedbe", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643906254-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac9a6686e7184fb1913e84a77985b449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape650d2e9-ac", "ovs_interfaceid": "e650d2e9-acd3-479b-9d48-748636ace077", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.401984] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356894, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.683057] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.243s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.683057] env[61906]: DEBUG nova.compute.manager [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 928.685644] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.755s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.687150] env[61906]: INFO nova.compute.claims [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 928.789773] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "refresh_cache-070f8882-9099-4f17-bae8-1701b5dcc08d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.790127] env[61906]: DEBUG nova.compute.manager [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Instance network_info: |[{"id": "e650d2e9-acd3-479b-9d48-748636ace077", "address": "fa:16:3e:71:a7:3c", "network": {"id": "5b29c9e0-23c9-42ee-9be4-60178bddedbe", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643906254-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac9a6686e7184fb1913e84a77985b449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape650d2e9-ac", "ovs_interfaceid": "e650d2e9-acd3-479b-9d48-748636ace077", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 928.790932] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:a7:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c02dd284-ab80-451c-93eb-48c8360acb9c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e650d2e9-acd3-479b-9d48-748636ace077', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.798451] env[61906]: DEBUG oslo.service.loopingcall [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.798700] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 928.799341] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c9d69043-dfa3-4cfb-a511-7f37af8dd9da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.817808] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.817808] env[61906]: value = "task-1356897" [ 928.817808] env[61906]: _type = "Task" [ 928.817808] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.825782] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356897, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.901714] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356894, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.440858} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.901949] env[61906]: INFO nova.virt.vmwareapi.ds_util [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_24a73825-6eed-45e1-a549-789d1098db8e/OSTACK_IMG_24a73825-6eed-45e1-a549-789d1098db8e.vmdk to [datastore1] devstack-image-cache_base/4e467722-4f47-4b26-9f34-13132817ea0e/4e467722-4f47-4b26-9f34-13132817ea0e.vmdk. [ 928.902157] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Cleaning up location [datastore1] OSTACK_IMG_24a73825-6eed-45e1-a549-789d1098db8e {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 928.902328] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_24a73825-6eed-45e1-a549-789d1098db8e {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 928.902581] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c967c59-d345-4da3-b5d8-30bb4d38d464 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.908602] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 928.908602] env[61906]: value = "task-1356898" [ 928.908602] env[61906]: _type = "Task" [ 928.908602] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.916158] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356898, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.028991] env[61906]: DEBUG nova.compute.manager [req-b8bfdaaa-7d9a-415a-b026-81a925b2a8bb req-ddd32199-8991-490d-b450-4d07240f1fc8 service nova] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Received event network-changed-e650d2e9-acd3-479b-9d48-748636ace077 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.029161] env[61906]: DEBUG nova.compute.manager [req-b8bfdaaa-7d9a-415a-b026-81a925b2a8bb req-ddd32199-8991-490d-b450-4d07240f1fc8 service nova] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Refreshing instance network info cache due to event network-changed-e650d2e9-acd3-479b-9d48-748636ace077. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 929.029388] env[61906]: DEBUG oslo_concurrency.lockutils [req-b8bfdaaa-7d9a-415a-b026-81a925b2a8bb req-ddd32199-8991-490d-b450-4d07240f1fc8 service nova] Acquiring lock "refresh_cache-070f8882-9099-4f17-bae8-1701b5dcc08d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.029538] env[61906]: DEBUG oslo_concurrency.lockutils [req-b8bfdaaa-7d9a-415a-b026-81a925b2a8bb req-ddd32199-8991-490d-b450-4d07240f1fc8 service nova] Acquired lock "refresh_cache-070f8882-9099-4f17-bae8-1701b5dcc08d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.029703] env[61906]: DEBUG nova.network.neutron [req-b8bfdaaa-7d9a-415a-b026-81a925b2a8bb req-ddd32199-8991-490d-b450-4d07240f1fc8 service nova] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Refreshing network info cache for port e650d2e9-acd3-479b-9d48-748636ace077 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 929.123180] env[61906]: DEBUG nova.compute.manager [req-f07b029b-244e-4deb-a52b-9bd4b51294fd req-efa02def-d90e-439b-84a0-92c79fd8780e service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Received event network-changed-eb78ba42-034c-44d7-9dd3-4157cd702c8c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.123180] env[61906]: DEBUG nova.compute.manager [req-f07b029b-244e-4deb-a52b-9bd4b51294fd req-efa02def-d90e-439b-84a0-92c79fd8780e service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Refreshing instance network info cache due to event network-changed-eb78ba42-034c-44d7-9dd3-4157cd702c8c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 929.123180] env[61906]: DEBUG oslo_concurrency.lockutils [req-f07b029b-244e-4deb-a52b-9bd4b51294fd req-efa02def-d90e-439b-84a0-92c79fd8780e service nova] Acquiring lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.123309] env[61906]: DEBUG oslo_concurrency.lockutils [req-f07b029b-244e-4deb-a52b-9bd4b51294fd req-efa02def-d90e-439b-84a0-92c79fd8780e service nova] Acquired lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.123385] env[61906]: DEBUG nova.network.neutron [req-f07b029b-244e-4deb-a52b-9bd4b51294fd req-efa02def-d90e-439b-84a0-92c79fd8780e service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Refreshing network info cache for port eb78ba42-034c-44d7-9dd3-4157cd702c8c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 929.192190] env[61906]: DEBUG nova.compute.utils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 929.195828] env[61906]: DEBUG nova.compute.manager [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Not allocating networking since 'none' was specified. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 929.327955] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356897, 'name': CreateVM_Task, 'duration_secs': 0.344441} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.329034] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 929.329034] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.329194] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.329387] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 929.329639] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd86d327-2828-473b-8551-a5510f5dde81 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.333849] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 929.333849] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52743ad5-353e-2ae0-76de-78b9540fb8da" [ 929.333849] env[61906]: _type = "Task" [ 929.333849] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.341511] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52743ad5-353e-2ae0-76de-78b9540fb8da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.418410] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356898, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.04258} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.418410] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.418753] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "[datastore1] devstack-image-cache_base/4e467722-4f47-4b26-9f34-13132817ea0e/4e467722-4f47-4b26-9f34-13132817ea0e.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.418753] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/4e467722-4f47-4b26-9f34-13132817ea0e/4e467722-4f47-4b26-9f34-13132817ea0e.vmdk to [datastore1] cb879dc0-3af7-4279-aa28-66e2b8b4286d/cb879dc0-3af7-4279-aa28-66e2b8b4286d.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 929.419022] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d93d9eb1-6eb8-43f8-bcd7-3bfca0942c61 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.425528] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 929.425528] env[61906]: value = "task-1356899" [ 929.425528] env[61906]: _type = "Task" [ 929.425528] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.433175] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356899, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.699884] env[61906]: DEBUG nova.compute.manager [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 929.844806] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52743ad5-353e-2ae0-76de-78b9540fb8da, 'name': SearchDatastore_Task, 'duration_secs': 0.013594} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.848261] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.848746] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.848746] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.848945] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.849102] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.851751] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-317ebeb1-61fb-4c2a-acb4-c5b1f6256d76 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.869223] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.869446] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 929.870246] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac88fb5c-f417-4fb0-839f-599569d078f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.876220] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 929.876220] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52711ba9-a76a-72fe-22c6-1b7fc0588403" [ 929.876220] env[61906]: _type = "Task" [ 929.876220] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.881950] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21472ea3-089e-4815-bf83-847b2e01de72 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.887963] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52711ba9-a76a-72fe-22c6-1b7fc0588403, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.893195] env[61906]: DEBUG nova.network.neutron [req-b8bfdaaa-7d9a-415a-b026-81a925b2a8bb req-ddd32199-8991-490d-b450-4d07240f1fc8 service nova] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Updated VIF entry in instance network info cache for port e650d2e9-acd3-479b-9d48-748636ace077. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 929.893744] env[61906]: DEBUG nova.network.neutron [req-b8bfdaaa-7d9a-415a-b026-81a925b2a8bb req-ddd32199-8991-490d-b450-4d07240f1fc8 service nova] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Updating instance_info_cache with network_info: [{"id": "e650d2e9-acd3-479b-9d48-748636ace077", "address": "fa:16:3e:71:a7:3c", "network": {"id": "5b29c9e0-23c9-42ee-9be4-60178bddedbe", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643906254-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ac9a6686e7184fb1913e84a77985b449", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape650d2e9-ac", "ovs_interfaceid": "e650d2e9-acd3-479b-9d48-748636ace077", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.895867] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ba07c1-211b-48ee-9015-0e560b99895f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.935424] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f19901f-82cf-47f7-9116-1bc128f4ca13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.946571] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356899, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.949839] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6776bf-fe04-4f88-b92a-8d763fcf3667 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.968558] env[61906]: DEBUG nova.compute.provider_tree [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.024173] env[61906]: DEBUG nova.network.neutron [req-f07b029b-244e-4deb-a52b-9bd4b51294fd req-efa02def-d90e-439b-84a0-92c79fd8780e service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updated VIF entry in instance network info cache for port eb78ba42-034c-44d7-9dd3-4157cd702c8c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 930.024689] env[61906]: DEBUG nova.network.neutron [req-f07b029b-244e-4deb-a52b-9bd4b51294fd req-efa02def-d90e-439b-84a0-92c79fd8780e service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updating instance_info_cache with network_info: [{"id": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "address": "fa:16:3e:f0:1b:68", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb78ba42-03", "ovs_interfaceid": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.388057] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52711ba9-a76a-72fe-22c6-1b7fc0588403, 'name': SearchDatastore_Task, 'duration_secs': 0.090107} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.388934] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36cb0a00-34f7-4ce2-9c77-07d5520273a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.396115] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 930.396115] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52835914-9f97-7cf7-ccef-39b7fd2d2654" [ 930.396115] env[61906]: _type = "Task" [ 930.396115] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.400270] env[61906]: DEBUG oslo_concurrency.lockutils [req-b8bfdaaa-7d9a-415a-b026-81a925b2a8bb req-ddd32199-8991-490d-b450-4d07240f1fc8 service nova] Releasing lock "refresh_cache-070f8882-9099-4f17-bae8-1701b5dcc08d" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.404949] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52835914-9f97-7cf7-ccef-39b7fd2d2654, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.441490] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356899, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.471814] env[61906]: DEBUG nova.scheduler.client.report [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.527419] env[61906]: DEBUG oslo_concurrency.lockutils [req-f07b029b-244e-4deb-a52b-9bd4b51294fd req-efa02def-d90e-439b-84a0-92c79fd8780e service nova] Releasing lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.711729] env[61906]: DEBUG nova.compute.manager [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 930.740198] env[61906]: DEBUG nova.virt.hardware [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.740517] env[61906]: DEBUG nova.virt.hardware [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.740745] env[61906]: DEBUG nova.virt.hardware [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.740949] env[61906]: DEBUG nova.virt.hardware [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.741122] env[61906]: DEBUG nova.virt.hardware [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.741308] env[61906]: DEBUG nova.virt.hardware [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.741583] env[61906]: DEBUG nova.virt.hardware [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.741763] env[61906]: DEBUG nova.virt.hardware [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.741939] env[61906]: DEBUG nova.virt.hardware [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.742121] env[61906]: DEBUG nova.virt.hardware [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.742304] env[61906]: DEBUG nova.virt.hardware [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.743210] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5faf6ae-885f-4849-9c34-ebf390c9a2b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.754452] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9c6f55-8dfd-4ff9-a846-0dc5bd34fc40 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.771379] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.777136] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Creating folder: Project (8d7878a309694c049413059cc17efe9a). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 930.777532] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4fd0b0e5-3c2a-4083-a325-7d2815f3058d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.789497] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Created folder: Project (8d7878a309694c049413059cc17efe9a) in parent group-v288914. [ 930.789722] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Creating folder: Instances. Parent ref: group-v289020. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 930.790021] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-55248fa5-5e90-4d04-8b5f-3cd33800c954 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.800223] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Created folder: Instances in parent group-v289020. [ 930.800489] env[61906]: DEBUG oslo.service.loopingcall [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.800716] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 930.800973] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6fee1fe1-f53d-4efd-937d-9f5f9027f659 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.819358] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.819358] env[61906]: value = "task-1356902" [ 930.819358] env[61906]: _type = "Task" [ 930.819358] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.828495] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356902, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.908605] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52835914-9f97-7cf7-ccef-39b7fd2d2654, 'name': SearchDatastore_Task, 'duration_secs': 0.086115} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.908858] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.909161] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 070f8882-9099-4f17-bae8-1701b5dcc08d/070f8882-9099-4f17-bae8-1701b5dcc08d.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 930.909442] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d1768451-fe1f-48b8-bbea-2f1f1ff2be13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.917377] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 930.917377] env[61906]: value = "task-1356903" [ 930.917377] env[61906]: _type = "Task" [ 930.917377] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.927749] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356903, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.942202] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356899, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.977338] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.292s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.977988] env[61906]: DEBUG nova.compute.manager [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 931.151290] env[61906]: DEBUG nova.compute.manager [req-26b75a1e-e274-4b25-a77a-766a07c7b2b5 req-794f4182-fcf4-435d-831d-279b36e4c0d5 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Received event network-changed-eb78ba42-034c-44d7-9dd3-4157cd702c8c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.151290] env[61906]: DEBUG nova.compute.manager [req-26b75a1e-e274-4b25-a77a-766a07c7b2b5 req-794f4182-fcf4-435d-831d-279b36e4c0d5 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Refreshing instance network info cache due to event network-changed-eb78ba42-034c-44d7-9dd3-4157cd702c8c. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 931.151700] env[61906]: DEBUG oslo_concurrency.lockutils [req-26b75a1e-e274-4b25-a77a-766a07c7b2b5 req-794f4182-fcf4-435d-831d-279b36e4c0d5 service nova] Acquiring lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.151700] env[61906]: DEBUG oslo_concurrency.lockutils [req-26b75a1e-e274-4b25-a77a-766a07c7b2b5 req-794f4182-fcf4-435d-831d-279b36e4c0d5 service nova] Acquired lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.151700] env[61906]: DEBUG nova.network.neutron [req-26b75a1e-e274-4b25-a77a-766a07c7b2b5 req-794f4182-fcf4-435d-831d-279b36e4c0d5 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Refreshing network info cache for port eb78ba42-034c-44d7-9dd3-4157cd702c8c {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 931.330662] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356902, 'name': CreateVM_Task, 'duration_secs': 0.379921} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.330895] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 931.331366] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.331559] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.331955] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 931.332255] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d3946e5-6f6d-45bb-a198-7cd54143fd19 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.338680] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 931.338680] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a9a95c-adc1-6978-4793-6927fdb33fb2" [ 931.338680] env[61906]: _type = "Task" [ 931.338680] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.349381] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a9a95c-adc1-6978-4793-6927fdb33fb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.429224] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356903, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.441656] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356899, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.483403] env[61906]: DEBUG nova.compute.utils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 931.485032] env[61906]: DEBUG nova.compute.manager [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 931.485305] env[61906]: DEBUG nova.network.neutron [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 931.533593] env[61906]: DEBUG nova.policy [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf4fea47b929461785f4a487a56977a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3093466de87a45a2ac673c25bea19bc3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 931.853175] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a9a95c-adc1-6978-4793-6927fdb33fb2, 'name': SearchDatastore_Task, 'duration_secs': 0.013662} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.853175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.853175] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.853175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.853175] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.853175] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.853175] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7134f991-3197-4f06-99bf-3af951187e17 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.862251] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.862918] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 931.863380] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c26a9931-db31-4c97-84ab-0bf02edda123 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.871733] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 931.871733] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52de1dc6-3d11-c4aa-60b3-957c0bba0278" [ 931.871733] env[61906]: _type = "Task" [ 931.871733] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.880787] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52de1dc6-3d11-c4aa-60b3-957c0bba0278, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.929033] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356903, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.943987] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356899, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.280128} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.944438] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/4e467722-4f47-4b26-9f34-13132817ea0e/4e467722-4f47-4b26-9f34-13132817ea0e.vmdk to [datastore1] cb879dc0-3af7-4279-aa28-66e2b8b4286d/cb879dc0-3af7-4279-aa28-66e2b8b4286d.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 931.945479] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8258348-87bd-4a05-8500-17470a87d62f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.970867] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] cb879dc0-3af7-4279-aa28-66e2b8b4286d/cb879dc0-3af7-4279-aa28-66e2b8b4286d.vmdk or device None with type streamOptimized {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.971284] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f67bde3-144b-4061-8043-72701c3e09e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.988137] env[61906]: DEBUG nova.network.neutron [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Successfully created port: ce54463c-a7f7-4c89-8d23-cf62086bfaa3 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 931.990356] env[61906]: DEBUG nova.compute.manager [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 931.999447] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 931.999447] env[61906]: value = "task-1356904" [ 931.999447] env[61906]: _type = "Task" [ 931.999447] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.009076] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356904, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.185413] env[61906]: DEBUG nova.network.neutron [req-26b75a1e-e274-4b25-a77a-766a07c7b2b5 req-794f4182-fcf4-435d-831d-279b36e4c0d5 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updated VIF entry in instance network info cache for port eb78ba42-034c-44d7-9dd3-4157cd702c8c. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 932.186013] env[61906]: DEBUG nova.network.neutron [req-26b75a1e-e274-4b25-a77a-766a07c7b2b5 req-794f4182-fcf4-435d-831d-279b36e4c0d5 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updating instance_info_cache with network_info: [{"id": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "address": "fa:16:3e:f0:1b:68", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb78ba42-03", "ovs_interfaceid": "eb78ba42-034c-44d7-9dd3-4157cd702c8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.382024] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52de1dc6-3d11-c4aa-60b3-957c0bba0278, 'name': SearchDatastore_Task, 'duration_secs': 0.017325} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.382827] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60de37b8-45ff-4c59-871e-6053ce76ae0e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.389027] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 932.389027] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d1a672-7cf7-bc66-60b4-1534668352a2" [ 932.389027] env[61906]: _type = "Task" [ 932.389027] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.396671] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d1a672-7cf7-bc66-60b4-1534668352a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.427022] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356903, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.145782} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.427243] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 070f8882-9099-4f17-bae8-1701b5dcc08d/070f8882-9099-4f17-bae8-1701b5dcc08d.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 932.427467] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.427751] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7fd3a42-b3d6-4d2c-9f23-469723fcef6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.433599] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 932.433599] env[61906]: value = "task-1356905" [ 932.433599] env[61906]: _type = "Task" [ 932.433599] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.440622] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356905, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.509080] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356904, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.689062] env[61906]: DEBUG oslo_concurrency.lockutils [req-26b75a1e-e274-4b25-a77a-766a07c7b2b5 req-794f4182-fcf4-435d-831d-279b36e4c0d5 service nova] Releasing lock "refresh_cache-7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.899531] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d1a672-7cf7-bc66-60b4-1534668352a2, 'name': SearchDatastore_Task, 'duration_secs': 0.011856} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.899822] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.900108] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] c77823f1-189c-4ac9-9a40-7bc45e82b442/c77823f1-189c-4ac9-9a40-7bc45e82b442.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 932.900358] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3274c8b0-a1c3-48b5-88d2-09c5cd5636a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.905717] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 932.905717] env[61906]: value = "task-1356906" [ 932.905717] env[61906]: _type = "Task" [ 932.905717] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.912617] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356906, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.944460] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356905, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063142} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.944740] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 932.945541] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e24a65c-3139-49a0-97d4-88385f3c9986 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.967916] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 070f8882-9099-4f17-bae8-1701b5dcc08d/070f8882-9099-4f17-bae8-1701b5dcc08d.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 932.968174] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c1427c5-166c-4392-a633-037ac2ce2524 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.987018] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 932.987018] env[61906]: value = "task-1356907" [ 932.987018] env[61906]: _type = "Task" [ 932.987018] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.998108] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356907, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.000270] env[61906]: DEBUG nova.compute.manager [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 933.010279] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356904, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.026545] env[61906]: DEBUG nova.virt.hardware [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.026785] env[61906]: DEBUG nova.virt.hardware [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.026955] env[61906]: DEBUG nova.virt.hardware [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.027186] env[61906]: DEBUG nova.virt.hardware [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.027315] env[61906]: DEBUG nova.virt.hardware [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.027472] env[61906]: DEBUG nova.virt.hardware [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.027805] env[61906]: DEBUG nova.virt.hardware [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.027988] env[61906]: DEBUG nova.virt.hardware [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.028179] env[61906]: DEBUG nova.virt.hardware [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.028348] env[61906]: DEBUG nova.virt.hardware [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.028530] env[61906]: DEBUG nova.virt.hardware [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.029349] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8844497-fb65-4f8e-be90-57f4aefa7a50 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.036695] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169671ac-3443-43ec-9c35-2fa3c8c710ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.415540] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356906, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.486159] env[61906]: DEBUG nova.compute.manager [req-5212ca12-34a2-4ee2-94ad-aad50b316dd2 req-53eec065-456c-4511-8024-a4c9419a585c service nova] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Received event network-vif-plugged-ce54463c-a7f7-4c89-8d23-cf62086bfaa3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.486394] env[61906]: DEBUG oslo_concurrency.lockutils [req-5212ca12-34a2-4ee2-94ad-aad50b316dd2 req-53eec065-456c-4511-8024-a4c9419a585c service nova] Acquiring lock "766e79cb-bea3-4b74-923d-a147c7be4134-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.486610] env[61906]: DEBUG oslo_concurrency.lockutils [req-5212ca12-34a2-4ee2-94ad-aad50b316dd2 req-53eec065-456c-4511-8024-a4c9419a585c service nova] Lock "766e79cb-bea3-4b74-923d-a147c7be4134-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.486862] env[61906]: DEBUG oslo_concurrency.lockutils [req-5212ca12-34a2-4ee2-94ad-aad50b316dd2 req-53eec065-456c-4511-8024-a4c9419a585c service nova] Lock "766e79cb-bea3-4b74-923d-a147c7be4134-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.486956] env[61906]: DEBUG nova.compute.manager [req-5212ca12-34a2-4ee2-94ad-aad50b316dd2 req-53eec065-456c-4511-8024-a4c9419a585c service nova] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] No waiting events found dispatching network-vif-plugged-ce54463c-a7f7-4c89-8d23-cf62086bfaa3 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 933.487305] env[61906]: WARNING nova.compute.manager [req-5212ca12-34a2-4ee2-94ad-aad50b316dd2 req-53eec065-456c-4511-8024-a4c9419a585c service nova] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Received unexpected event network-vif-plugged-ce54463c-a7f7-4c89-8d23-cf62086bfaa3 for instance with vm_state building and task_state spawning. [ 933.497017] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356907, 'name': ReconfigVM_Task, 'duration_secs': 0.307221} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.497303] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 070f8882-9099-4f17-bae8-1701b5dcc08d/070f8882-9099-4f17-bae8-1701b5dcc08d.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.497970] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f63ff692-e35f-4272-8037-887477d69617 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.504243] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 933.504243] env[61906]: value = "task-1356908" [ 933.504243] env[61906]: _type = "Task" [ 933.504243] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.514911] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356904, 'name': ReconfigVM_Task, 'duration_secs': 1.215419} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.517909] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Reconfigured VM instance instance-00000049 to attach disk [datastore1] cb879dc0-3af7-4279-aa28-66e2b8b4286d/cb879dc0-3af7-4279-aa28-66e2b8b4286d.vmdk or device None with type streamOptimized {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.519277] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'size': 0, 'device_name': '/dev/sda', 'encryption_format': None, 'encryption_options': None, 'disk_bus': None, 'guest_format': None, 'encrypted': False, 'encryption_secret_uuid': None, 'device_type': 'disk', 'boot_index': 0, 'image_id': '29664dd5-796a-4dd3-b9f0-6a185cb0b7c1'}], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'mount_device': '/dev/sdb', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-289016', 'volume_id': '4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'name': 'volume-4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'cb879dc0-3af7-4279-aa28-66e2b8b4286d', 'attached_at': '', 'detached_at': '', 'volume_id': '4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'serial': '4d5078bb-42a4-4eaa-b253-4e94f1321e51'}, 'attachment_id': '7b4d0a0d-6cf1-48d8-b2a1-67c7f8cbc6a0', 'guest_format': None, 'delete_on_termination': False, 'disk_bus': None, 'boot_index': None, 'volume_type': None}], 'swap': None} {{(pid=61906) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 933.519491] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 933.519682] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-289016', 'volume_id': '4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'name': 'volume-4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'cb879dc0-3af7-4279-aa28-66e2b8b4286d', 'attached_at': '', 'detached_at': '', 'volume_id': '4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'serial': '4d5078bb-42a4-4eaa-b253-4e94f1321e51'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 933.520010] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356908, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.520702] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61429614-4050-4fee-b389-57f257ddb924 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.535398] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-810cc52a-ba99-4061-8e44-9c6900d3bb54 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.559812] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] volume-4d5078bb-42a4-4eaa-b253-4e94f1321e51/volume-4d5078bb-42a4-4eaa-b253-4e94f1321e51.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.560490] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f61e2178-cf8c-4e24-8e69-7d2a1a4a026e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.576308] env[61906]: DEBUG nova.network.neutron [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Successfully updated port: ce54463c-a7f7-4c89-8d23-cf62086bfaa3 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 933.579620] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 933.579620] env[61906]: value = "task-1356909" [ 933.579620] env[61906]: _type = "Task" [ 933.579620] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.591621] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356909, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.915912] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356906, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.554884} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.916196] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] c77823f1-189c-4ac9-9a40-7bc45e82b442/c77823f1-189c-4ac9-9a40-7bc45e82b442.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 933.916415] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 933.917051] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5e007957-b0e1-4616-9000-6e2c66cb0d73 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.922949] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 933.922949] env[61906]: value = "task-1356910" [ 933.922949] env[61906]: _type = "Task" [ 933.922949] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.931214] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356910, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.017236] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356908, 'name': Rename_Task, 'duration_secs': 0.131643} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.017658] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 934.017978] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6cb7d10c-927e-433b-aef5-3bcbe5ac7193 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.025608] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 934.025608] env[61906]: value = "task-1356911" [ 934.025608] env[61906]: _type = "Task" [ 934.025608] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.033547] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.079423] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "refresh_cache-766e79cb-bea3-4b74-923d-a147c7be4134" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.079578] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "refresh_cache-766e79cb-bea3-4b74-923d-a147c7be4134" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.079794] env[61906]: DEBUG nova.network.neutron [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 934.089851] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356909, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.433619] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356910, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068403} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.434078] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 934.434960] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca5a85a-2546-4b2f-9b69-fbf8f96b8178 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.454392] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] c77823f1-189c-4ac9-9a40-7bc45e82b442/c77823f1-189c-4ac9-9a40-7bc45e82b442.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.454652] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b36cdd98-cb81-4b52-ae63-cb8d5803e221 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.474499] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 934.474499] env[61906]: value = "task-1356912" [ 934.474499] env[61906]: _type = "Task" [ 934.474499] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.482578] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356912, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.537196] env[61906]: DEBUG oslo_vmware.api [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356911, 'name': PowerOnVM_Task, 'duration_secs': 0.489661} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.537545] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 934.538373] env[61906]: INFO nova.compute.manager [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Took 7.84 seconds to spawn the instance on the hypervisor. [ 934.538373] env[61906]: DEBUG nova.compute.manager [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.539055] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f2fa7d-af6d-4b10-8957-99932670d33c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.593706] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356909, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.614422] env[61906]: DEBUG nova.network.neutron [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 934.762218] env[61906]: DEBUG nova.network.neutron [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Updating instance_info_cache with network_info: [{"id": "ce54463c-a7f7-4c89-8d23-cf62086bfaa3", "address": "fa:16:3e:f3:a5:a5", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce54463c-a7", "ovs_interfaceid": "ce54463c-a7f7-4c89-8d23-cf62086bfaa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.985977] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356912, 'name': ReconfigVM_Task, 'duration_secs': 0.309622} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.986901] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Reconfigured VM instance instance-0000005b to attach disk [datastore2] c77823f1-189c-4ac9-9a40-7bc45e82b442/c77823f1-189c-4ac9-9a40-7bc45e82b442.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 934.987050] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-73a3a9a3-718d-41db-8ecb-4d4df8fe63a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.994651] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 934.994651] env[61906]: value = "task-1356913" [ 934.994651] env[61906]: _type = "Task" [ 934.994651] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.002477] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356913, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.056203] env[61906]: INFO nova.compute.manager [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Took 12.97 seconds to build instance. [ 935.092939] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356909, 'name': ReconfigVM_Task, 'duration_secs': 1.099763} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.092939] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Reconfigured VM instance instance-00000049 to attach disk [datastore2] volume-4d5078bb-42a4-4eaa-b253-4e94f1321e51/volume-4d5078bb-42a4-4eaa-b253-4e94f1321e51.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.097713] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5ec50ab-6bd5-4b48-a6ca-00292f72110b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.114354] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 935.114354] env[61906]: value = "task-1356914" [ 935.114354] env[61906]: _type = "Task" [ 935.114354] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.122737] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356914, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.265191] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "refresh_cache-766e79cb-bea3-4b74-923d-a147c7be4134" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.265965] env[61906]: DEBUG nova.compute.manager [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Instance network_info: |[{"id": "ce54463c-a7f7-4c89-8d23-cf62086bfaa3", "address": "fa:16:3e:f3:a5:a5", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce54463c-a7", "ovs_interfaceid": "ce54463c-a7f7-4c89-8d23-cf62086bfaa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 935.265965] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:a5:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30f1dacf-8988-41b8-aa8f-e9530f65ef46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ce54463c-a7f7-4c89-8d23-cf62086bfaa3', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 935.273681] env[61906]: DEBUG oslo.service.loopingcall [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.273930] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 935.274193] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e59e9fb3-0bf0-434f-8e6e-54bffea2e908 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.297031] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 935.297031] env[61906]: value = "task-1356915" [ 935.297031] env[61906]: _type = "Task" [ 935.297031] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.305592] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356915, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.505659] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356913, 'name': Rename_Task, 'duration_secs': 0.140441} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.505943] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 935.506172] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ebaf5f0f-1831-4b74-aa10-a7fcea34941a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.512312] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 935.512312] env[61906]: value = "task-1356916" [ 935.512312] env[61906]: _type = "Task" [ 935.512312] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.520224] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356916, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.558350] env[61906]: DEBUG oslo_concurrency.lockutils [None req-817e9df1-92e5-409f-8279-43a8ba4f6fff tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "070f8882-9099-4f17-bae8-1701b5dcc08d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.489s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.604716] env[61906]: DEBUG nova.compute.manager [req-0145ce78-aa9c-4f76-bb87-49244b2ba14e req-1ca64254-f8d4-4037-8865-e0e14bc29a29 service nova] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Received event network-changed-ce54463c-a7f7-4c89-8d23-cf62086bfaa3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.604987] env[61906]: DEBUG nova.compute.manager [req-0145ce78-aa9c-4f76-bb87-49244b2ba14e req-1ca64254-f8d4-4037-8865-e0e14bc29a29 service nova] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Refreshing instance network info cache due to event network-changed-ce54463c-a7f7-4c89-8d23-cf62086bfaa3. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 935.605238] env[61906]: DEBUG oslo_concurrency.lockutils [req-0145ce78-aa9c-4f76-bb87-49244b2ba14e req-1ca64254-f8d4-4037-8865-e0e14bc29a29 service nova] Acquiring lock "refresh_cache-766e79cb-bea3-4b74-923d-a147c7be4134" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.605392] env[61906]: DEBUG oslo_concurrency.lockutils [req-0145ce78-aa9c-4f76-bb87-49244b2ba14e req-1ca64254-f8d4-4037-8865-e0e14bc29a29 service nova] Acquired lock "refresh_cache-766e79cb-bea3-4b74-923d-a147c7be4134" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.605562] env[61906]: DEBUG nova.network.neutron [req-0145ce78-aa9c-4f76-bb87-49244b2ba14e req-1ca64254-f8d4-4037-8865-e0e14bc29a29 service nova] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Refreshing network info cache for port ce54463c-a7f7-4c89-8d23-cf62086bfaa3 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 935.624729] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356914, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.793370] env[61906]: DEBUG oslo_concurrency.lockutils [None req-84e4b7fb-b838-4245-b7ef-ba0957d9cc93 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "interface-99e5c55d-79ae-47a6-8500-79fc68291650-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.793931] env[61906]: DEBUG oslo_concurrency.lockutils [None req-84e4b7fb-b838-4245-b7ef-ba0957d9cc93 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-99e5c55d-79ae-47a6-8500-79fc68291650-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.793931] env[61906]: DEBUG nova.objects.instance [None req-84e4b7fb-b838-4245-b7ef-ba0957d9cc93 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'flavor' on Instance uuid 99e5c55d-79ae-47a6-8500-79fc68291650 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.807817] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356915, 'name': CreateVM_Task, 'duration_secs': 0.327426} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.808549] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 935.809230] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.809400] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.809717] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 935.810211] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5469c3e7-f273-4a12-b0fa-ee57cff99b18 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.814615] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 935.814615] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528e3104-bc95-6841-a705-57e1624e9606" [ 935.814615] env[61906]: _type = "Task" [ 935.814615] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.822314] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528e3104-bc95-6841-a705-57e1624e9606, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.022848] env[61906]: DEBUG oslo_vmware.api [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356916, 'name': PowerOnVM_Task, 'duration_secs': 0.451787} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.023857] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 936.023857] env[61906]: INFO nova.compute.manager [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Took 5.31 seconds to spawn the instance on the hypervisor. [ 936.023857] env[61906]: DEBUG nova.compute.manager [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 936.024512] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46bbca0a-2453-484a-939a-d200354e17a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.123678] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356914, 'name': ReconfigVM_Task, 'duration_secs': 0.657836} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.124327] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-289016', 'volume_id': '4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'name': 'volume-4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'cb879dc0-3af7-4279-aa28-66e2b8b4286d', 'attached_at': '', 'detached_at': '', 'volume_id': '4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'serial': '4d5078bb-42a4-4eaa-b253-4e94f1321e51'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 936.124899] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-59a35071-b9c7-43b4-b297-e4cd9e782d99 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.131620] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 936.131620] env[61906]: value = "task-1356917" [ 936.131620] env[61906]: _type = "Task" [ 936.131620] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.140454] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356917, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.297565] env[61906]: DEBUG nova.objects.instance [None req-84e4b7fb-b838-4245-b7ef-ba0957d9cc93 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'pci_requests' on Instance uuid 99e5c55d-79ae-47a6-8500-79fc68291650 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.306557] env[61906]: DEBUG nova.network.neutron [req-0145ce78-aa9c-4f76-bb87-49244b2ba14e req-1ca64254-f8d4-4037-8865-e0e14bc29a29 service nova] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Updated VIF entry in instance network info cache for port ce54463c-a7f7-4c89-8d23-cf62086bfaa3. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 936.306921] env[61906]: DEBUG nova.network.neutron [req-0145ce78-aa9c-4f76-bb87-49244b2ba14e req-1ca64254-f8d4-4037-8865-e0e14bc29a29 service nova] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Updating instance_info_cache with network_info: [{"id": "ce54463c-a7f7-4c89-8d23-cf62086bfaa3", "address": "fa:16:3e:f3:a5:a5", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce54463c-a7", "ovs_interfaceid": "ce54463c-a7f7-4c89-8d23-cf62086bfaa3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.324782] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528e3104-bc95-6841-a705-57e1624e9606, 'name': SearchDatastore_Task, 'duration_secs': 0.009746} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.325093] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.325327] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 936.325565] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.325718] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.325903] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.326185] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac637b0a-83de-42f6-a4b2-368a948bf177 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.334321] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.334499] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 936.335206] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bc79b59-65fb-4306-8d18-b7f6adab54e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.340554] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 936.340554] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5251b6a5-7ffc-0571-d04f-1dd27d2be4db" [ 936.340554] env[61906]: _type = "Task" [ 936.340554] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.348131] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5251b6a5-7ffc-0571-d04f-1dd27d2be4db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.444688] env[61906]: DEBUG nova.compute.manager [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 936.445616] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca50312-30e0-4f6d-b8d1-6c4e2f821994 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.539511] env[61906]: INFO nova.compute.manager [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Took 11.86 seconds to build instance. [ 936.641906] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356917, 'name': Rename_Task, 'duration_secs': 0.303809} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.642247] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 936.642511] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ec00457-dec7-445d-88c4-b832b727df1a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.648939] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 936.648939] env[61906]: value = "task-1356918" [ 936.648939] env[61906]: _type = "Task" [ 936.648939] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.657944] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356918, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.800715] env[61906]: DEBUG nova.objects.base [None req-84e4b7fb-b838-4245-b7ef-ba0957d9cc93 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Object Instance<99e5c55d-79ae-47a6-8500-79fc68291650> lazy-loaded attributes: flavor,pci_requests {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 936.800919] env[61906]: DEBUG nova.network.neutron [None req-84e4b7fb-b838-4245-b7ef-ba0957d9cc93 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 936.809523] env[61906]: DEBUG oslo_concurrency.lockutils [req-0145ce78-aa9c-4f76-bb87-49244b2ba14e req-1ca64254-f8d4-4037-8865-e0e14bc29a29 service nova] Releasing lock "refresh_cache-766e79cb-bea3-4b74-923d-a147c7be4134" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.850760] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5251b6a5-7ffc-0571-d04f-1dd27d2be4db, 'name': SearchDatastore_Task, 'duration_secs': 0.008762} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.851562] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b09ab557-4dfb-4f4d-a662-bbddaf9f7af4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.856675] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 936.856675] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529d5f9b-7c00-1579-6f19-10137b34072c" [ 936.856675] env[61906]: _type = "Task" [ 936.856675] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.864276] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529d5f9b-7c00-1579-6f19-10137b34072c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.915752] env[61906]: DEBUG oslo_concurrency.lockutils [None req-84e4b7fb-b838-4245-b7ef-ba0957d9cc93 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-99e5c55d-79ae-47a6-8500-79fc68291650-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.122s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.956239] env[61906]: INFO nova.compute.manager [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] instance snapshotting [ 936.959073] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0ec2e6-0274-4826-a246-60775dc4956f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.977422] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e20d251-8da9-412c-a0fb-46a9edefce45 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.041499] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c1323f8f-718c-4ac3-9feb-20b183c82123 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Lock "c77823f1-189c-4ac9-9a40-7bc45e82b442" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.370s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.159934] env[61906]: DEBUG oslo_vmware.api [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356918, 'name': PowerOnVM_Task, 'duration_secs': 0.495398} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.160415] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 937.252986] env[61906]: DEBUG nova.compute.manager [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.253936] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77838710-21a4-4ae1-8bcf-358621e7adbd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.367272] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529d5f9b-7c00-1579-6f19-10137b34072c, 'name': SearchDatastore_Task, 'duration_secs': 0.010082} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.367557] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.367852] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 766e79cb-bea3-4b74-923d-a147c7be4134/766e79cb-bea3-4b74-923d-a147c7be4134.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 937.368129] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7180a8f9-5e30-42d1-b961-85a0e814f6bf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.374069] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 937.374069] env[61906]: value = "task-1356919" [ 937.374069] env[61906]: _type = "Task" [ 937.374069] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.381978] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356919, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.487751] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 937.488085] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-674aee65-3b07-453b-9803-c621baf35c41 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.497654] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 937.497654] env[61906]: value = "task-1356920" [ 937.497654] env[61906]: _type = "Task" [ 937.497654] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.505958] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356920, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.769043] env[61906]: INFO nova.compute.manager [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Rebuilding instance [ 937.779288] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e0c689af-951a-46b8-9787-2100e8562a6c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 31.415s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.824627] env[61906]: DEBUG nova.compute.manager [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.825651] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69267f12-245b-4e13-841e-cebf8d2949e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.885420] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356919, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.007827] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356920, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.339078] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 938.340087] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-16af9fea-3ec7-4631-8b1c-c6f1ec1c2047 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.346795] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 938.346795] env[61906]: value = "task-1356921" [ 938.346795] env[61906]: _type = "Task" [ 938.346795] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.355739] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356921, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.384280] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356919, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.580939} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.384572] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 766e79cb-bea3-4b74-923d-a147c7be4134/766e79cb-bea3-4b74-923d-a147c7be4134.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 938.384787] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 938.385220] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-33a2a80f-74f9-408b-bc22-45bc1d3973eb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.391395] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 938.391395] env[61906]: value = "task-1356922" [ 938.391395] env[61906]: _type = "Task" [ 938.391395] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.399975] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356922, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.507497] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356920, 'name': CreateSnapshot_Task, 'duration_secs': 0.825088} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.507778] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 938.508502] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320b4b9b-aa47-4457-a36b-6dd912f10e66 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.705386] env[61906]: DEBUG oslo_concurrency.lockutils [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.705649] env[61906]: DEBUG oslo_concurrency.lockutils [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.827751] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "interface-99e5c55d-79ae-47a6-8500-79fc68291650-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.828046] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-99e5c55d-79ae-47a6-8500-79fc68291650-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.828378] env[61906]: DEBUG nova.objects.instance [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'flavor' on Instance uuid 99e5c55d-79ae-47a6-8500-79fc68291650 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.857268] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356921, 'name': PowerOffVM_Task, 'duration_secs': 0.112668} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.857535] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 938.857793] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.858542] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c49b6408-705f-40c0-bfca-fdc718140af0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.864964] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.865204] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f1beb6de-06f8-4099-906a-328d856b8a09 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.889815] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.890077] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.890251] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Deleting the datastore file [datastore2] c77823f1-189c-4ac9-9a40-7bc45e82b442 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.890490] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4315a436-4961-459f-b5ca-3777dc9353c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.899976] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356922, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062727} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.901215] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 938.901594] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 938.901594] env[61906]: value = "task-1356924" [ 938.901594] env[61906]: _type = "Task" [ 938.901594] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.902426] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7afd6959-f770-470d-955e-3d954ccb34ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.926431] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 766e79cb-bea3-4b74-923d-a147c7be4134/766e79cb-bea3-4b74-923d-a147c7be4134.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 938.929308] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3c63f37-387c-4a20-b356-f31353d93a7b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.942969] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356924, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.948159] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 938.948159] env[61906]: value = "task-1356925" [ 938.948159] env[61906]: _type = "Task" [ 938.948159] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.960896] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356925, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.024777] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 939.025137] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-456a96cd-bd38-4f16-b4c5-7e505c60f62c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.032379] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 939.032379] env[61906]: value = "task-1356926" [ 939.032379] env[61906]: _type = "Task" [ 939.032379] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.040289] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356926, 'name': CloneVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.211090] env[61906]: DEBUG nova.compute.utils [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 939.411116] env[61906]: DEBUG nova.objects.instance [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'pci_requests' on Instance uuid 99e5c55d-79ae-47a6-8500-79fc68291650 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.415194] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356924, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121105} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.415651] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.415844] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 939.416039] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 939.457748] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356925, 'name': ReconfigVM_Task, 'duration_secs': 0.26742} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.458095] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 766e79cb-bea3-4b74-923d-a147c7be4134/766e79cb-bea3-4b74-923d-a147c7be4134.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 939.458768] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4db65265-0978-43f5-9c64-ebfe36af0256 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.464802] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 939.464802] env[61906]: value = "task-1356927" [ 939.464802] env[61906]: _type = "Task" [ 939.464802] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.474744] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356927, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.542488] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356926, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.714772] env[61906]: DEBUG oslo_concurrency.lockutils [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.916975] env[61906]: DEBUG nova.objects.base [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Object Instance<99e5c55d-79ae-47a6-8500-79fc68291650> lazy-loaded attributes: flavor,pci_requests {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 939.917366] env[61906]: DEBUG nova.network.neutron [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 939.961629] env[61906]: DEBUG nova.policy [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73219f5e96f64d1da86a19da9b18b06a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5533be74874b4094b7ef0f98a121f60b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 939.974292] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356927, 'name': Rename_Task, 'duration_secs': 0.127345} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.974578] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 939.974844] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e189902c-b809-44c3-b23d-e720f6ce644b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.980865] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 939.980865] env[61906]: value = "task-1356928" [ 939.980865] env[61906]: _type = "Task" [ 939.980865] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.988835] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356928, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.042235] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356926, 'name': CloneVM_Task} progress is 95%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.252747] env[61906]: DEBUG nova.network.neutron [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Successfully created port: f9078c39-0422-4f10-8f5e-12653d99d0ac {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 940.459172] env[61906]: DEBUG nova.virt.hardware [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.459464] env[61906]: DEBUG nova.virt.hardware [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.459630] env[61906]: DEBUG nova.virt.hardware [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.459820] env[61906]: DEBUG nova.virt.hardware [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.459977] env[61906]: DEBUG nova.virt.hardware [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.460210] env[61906]: DEBUG nova.virt.hardware [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.460436] env[61906]: DEBUG nova.virt.hardware [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.460599] env[61906]: DEBUG nova.virt.hardware [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.460773] env[61906]: DEBUG nova.virt.hardware [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.460942] env[61906]: DEBUG nova.virt.hardware [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.461133] env[61906]: DEBUG nova.virt.hardware [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.462016] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0de891-828f-4cec-bf94-a8625ac4c9c2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.469759] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7025a16-c320-4323-8c31-1559103fd0e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.483715] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Instance VIF info [] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.489729] env[61906]: DEBUG oslo.service.loopingcall [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.492775] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 940.493027] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c8f1060-01a5-431f-990b-563984b40c6f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.510494] env[61906]: DEBUG oslo_vmware.api [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356928, 'name': PowerOnVM_Task, 'duration_secs': 0.432709} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.511778] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 940.512000] env[61906]: INFO nova.compute.manager [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Took 7.51 seconds to spawn the instance on the hypervisor. [ 940.512202] env[61906]: DEBUG nova.compute.manager [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.512454] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.512454] env[61906]: value = "task-1356929" [ 940.512454] env[61906]: _type = "Task" [ 940.512454] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.513168] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b70934-0929-4f1c-86bb-73d5eefeab80 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.526915] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356929, 'name': CreateVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.541652] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356926, 'name': CloneVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.787495] env[61906]: DEBUG oslo_concurrency.lockutils [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.787881] env[61906]: DEBUG oslo_concurrency.lockutils [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.788166] env[61906]: INFO nova.compute.manager [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Attaching volume 04f4539c-399e-4b60-be5b-43e8125b1e8a to /dev/sdb [ 940.820064] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9b01c0-982c-456e-9867-da397feb959d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.830051] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a5813d-246c-4998-86fb-89b8af44dc01 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.843951] env[61906]: DEBUG nova.virt.block_device [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Updating existing volume attachment record: 0ea5231c-9b3c-4394-808b-b3284c62a4b1 {{(pid=61906) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 941.025590] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356929, 'name': CreateVM_Task} progress is 25%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.037730] env[61906]: INFO nova.compute.manager [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Took 16.12 seconds to build instance. [ 941.043982] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356926, 'name': CloneVM_Task, 'duration_secs': 1.511361} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.044275] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Created linked-clone VM from snapshot [ 941.045037] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613ebfc3-abde-4080-8a9c-bbb70683e7ed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.052440] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Uploading image 9a44efd7-e306-40d5-a03d-3f4b92997aba {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 941.063496] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 941.063791] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-36a8814c-fd43-49cf-b1ba-d5b622d39bbf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.070306] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 941.070306] env[61906]: value = "task-1356931" [ 941.070306] env[61906]: _type = "Task" [ 941.070306] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.079287] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356931, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.526407] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356929, 'name': CreateVM_Task, 'duration_secs': 0.673471} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.526407] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 941.526862] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.527050] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.527391] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.527644] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78780c5b-7f89-4bcb-bea0-63e8c773ab0a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.531878] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 941.531878] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ef1e4f-d940-6f79-10e8-a0738144380a" [ 941.531878] env[61906]: _type = "Task" [ 941.531878] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.539517] env[61906]: DEBUG oslo_concurrency.lockutils [None req-edb3e53b-7b4c-4cf6-afb8-a246811d2cae tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "766e79cb-bea3-4b74-923d-a147c7be4134" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.630s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.539752] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ef1e4f-d940-6f79-10e8-a0738144380a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.580780] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356931, 'name': Destroy_Task} progress is 33%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.769111] env[61906]: DEBUG nova.network.neutron [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Successfully updated port: f9078c39-0422-4f10-8f5e-12653d99d0ac {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 941.854510] env[61906]: DEBUG nova.compute.manager [req-11ff935c-ecdd-4b0b-8936-a2b453f1164d req-eb7d120c-5706-46b5-bdac-fc638b5261e6 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Received event network-vif-plugged-f9078c39-0422-4f10-8f5e-12653d99d0ac {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.854510] env[61906]: DEBUG oslo_concurrency.lockutils [req-11ff935c-ecdd-4b0b-8936-a2b453f1164d req-eb7d120c-5706-46b5-bdac-fc638b5261e6 service nova] Acquiring lock "99e5c55d-79ae-47a6-8500-79fc68291650-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.854510] env[61906]: DEBUG oslo_concurrency.lockutils [req-11ff935c-ecdd-4b0b-8936-a2b453f1164d req-eb7d120c-5706-46b5-bdac-fc638b5261e6 service nova] Lock "99e5c55d-79ae-47a6-8500-79fc68291650-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.854510] env[61906]: DEBUG oslo_concurrency.lockutils [req-11ff935c-ecdd-4b0b-8936-a2b453f1164d req-eb7d120c-5706-46b5-bdac-fc638b5261e6 service nova] Lock "99e5c55d-79ae-47a6-8500-79fc68291650-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.854890] env[61906]: DEBUG nova.compute.manager [req-11ff935c-ecdd-4b0b-8936-a2b453f1164d req-eb7d120c-5706-46b5-bdac-fc638b5261e6 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] No waiting events found dispatching network-vif-plugged-f9078c39-0422-4f10-8f5e-12653d99d0ac {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 941.854890] env[61906]: WARNING nova.compute.manager [req-11ff935c-ecdd-4b0b-8936-a2b453f1164d req-eb7d120c-5706-46b5-bdac-fc638b5261e6 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Received unexpected event network-vif-plugged-f9078c39-0422-4f10-8f5e-12653d99d0ac for instance with vm_state active and task_state None. [ 942.042442] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ef1e4f-d940-6f79-10e8-a0738144380a, 'name': SearchDatastore_Task, 'duration_secs': 0.009376} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.042826] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.042996] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 942.043255] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.043407] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.043594] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 942.043856] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c97e9366-a890-4235-91d2-7141497b8aa2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.052053] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 942.052280] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 942.052972] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5fa73c4-2481-4371-97eb-71849a2e54a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.059409] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 942.059409] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52315883-c117-9827-da47-cc5181dfccab" [ 942.059409] env[61906]: _type = "Task" [ 942.059409] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.066418] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52315883-c117-9827-da47-cc5181dfccab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.079354] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356931, 'name': Destroy_Task, 'duration_secs': 0.627928} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.079547] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Destroyed the VM [ 942.079779] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 942.080018] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-dbac389b-9832-4c9a-8815-b718b11c8517 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.085513] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 942.085513] env[61906]: value = "task-1356932" [ 942.085513] env[61906]: _type = "Task" [ 942.085513] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.091452] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f33a4e-ca36-415f-821b-18065430c2a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.096345] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356932, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.099332] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-155d27ee-616f-4e98-94d1-2dc156da83a1 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Suspending the VM {{(pid=61906) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 942.099546] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-833ac8f6-33b1-4086-a505-a74a2d70b7f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.104751] env[61906]: DEBUG oslo_vmware.api [None req-155d27ee-616f-4e98-94d1-2dc156da83a1 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 942.104751] env[61906]: value = "task-1356933" [ 942.104751] env[61906]: _type = "Task" [ 942.104751] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.111860] env[61906]: DEBUG oslo_vmware.api [None req-155d27ee-616f-4e98-94d1-2dc156da83a1 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356933, 'name': SuspendVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.273871] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.273871] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.273871] env[61906]: DEBUG nova.network.neutron [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 942.570872] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52315883-c117-9827-da47-cc5181dfccab, 'name': SearchDatastore_Task, 'duration_secs': 0.015818} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.571767] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8577c45-1428-47b2-8da6-33c5811a773a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.577103] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 942.577103] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fd87f3-63ee-4330-4019-f6724685c5de" [ 942.577103] env[61906]: _type = "Task" [ 942.577103] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.584881] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fd87f3-63ee-4330-4019-f6724685c5de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.593616] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356932, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.613812] env[61906]: DEBUG oslo_vmware.api [None req-155d27ee-616f-4e98-94d1-2dc156da83a1 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356933, 'name': SuspendVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.809916] env[61906]: WARNING nova.network.neutron [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] b12c63e1-bd61-4da3-8d37-4a89206ba0f2 already exists in list: networks containing: ['b12c63e1-bd61-4da3-8d37-4a89206ba0f2']. ignoring it [ 942.953205] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 942.953442] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 943.087444] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fd87f3-63ee-4330-4019-f6724685c5de, 'name': SearchDatastore_Task, 'duration_secs': 0.010134} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.093230] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.093569] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] c77823f1-189c-4ac9-9a40-7bc45e82b442/c77823f1-189c-4ac9-9a40-7bc45e82b442.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 943.093890] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a3809fd-4d46-49b4-881e-6bdc39b0c898 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.100763] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356932, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.102085] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 943.102085] env[61906]: value = "task-1356935" [ 943.102085] env[61906]: _type = "Task" [ 943.102085] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.113421] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356935, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.118128] env[61906]: DEBUG oslo_vmware.api [None req-155d27ee-616f-4e98-94d1-2dc156da83a1 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356933, 'name': SuspendVM_Task, 'duration_secs': 0.742806} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.118408] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-155d27ee-616f-4e98-94d1-2dc156da83a1 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Suspended the VM {{(pid=61906) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 943.118632] env[61906]: DEBUG nova.compute.manager [None req-155d27ee-616f-4e98-94d1-2dc156da83a1 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.119456] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f97cc4-4ead-4bb6-8109-8e4f3ab50e6c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.130757] env[61906]: DEBUG nova.network.neutron [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updating instance_info_cache with network_info: [{"id": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "address": "fa:16:3e:b8:da:a6", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a08ec03-40", "ovs_interfaceid": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f9078c39-0422-4f10-8f5e-12653d99d0ac", "address": "fa:16:3e:57:10:a2", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9078c39-04", "ovs_interfaceid": "f9078c39-0422-4f10-8f5e-12653d99d0ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.464053] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 943.464053] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 943.464053] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 943.596389] env[61906]: DEBUG oslo_vmware.api [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356932, 'name': RemoveSnapshot_Task, 'duration_secs': 1.384852} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.596674] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 943.609883] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356935, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.414779} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.610147] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] c77823f1-189c-4ac9-9a40-7bc45e82b442/c77823f1-189c-4ac9-9a40-7bc45e82b442.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 943.610371] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 943.610606] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b6e19fb-2ec7-4527-90f8-6811701f0062 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.616376] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 943.616376] env[61906]: value = "task-1356936" [ 943.616376] env[61906]: _type = "Task" [ 943.616376] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.623102] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356936, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.633467] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.634068] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.634244] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.635221] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd5c3c3-07af-43f7-905f-525bdafd5689 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.653264] env[61906]: DEBUG nova.virt.hardware [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 943.653456] env[61906]: DEBUG nova.virt.hardware [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 943.653621] env[61906]: DEBUG nova.virt.hardware [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 943.653812] env[61906]: DEBUG nova.virt.hardware [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 943.653965] env[61906]: DEBUG nova.virt.hardware [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 943.654135] env[61906]: DEBUG nova.virt.hardware [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 943.654347] env[61906]: DEBUG nova.virt.hardware [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 943.654507] env[61906]: DEBUG nova.virt.hardware [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 943.654676] env[61906]: DEBUG nova.virt.hardware [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 943.654846] env[61906]: DEBUG nova.virt.hardware [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 943.655043] env[61906]: DEBUG nova.virt.hardware [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 943.661190] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Reconfiguring VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 943.661723] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be2028d8-d323-4866-902d-0df4bdc36963 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.678704] env[61906]: DEBUG oslo_vmware.api [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 943.678704] env[61906]: value = "task-1356937" [ 943.678704] env[61906]: _type = "Task" [ 943.678704] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.688192] env[61906]: DEBUG oslo_vmware.api [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356937, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.929184] env[61906]: DEBUG nova.compute.manager [req-f4be150e-5729-4bfe-9043-564407421cbc req-977b3865-5c22-4d9a-93d7-91027e61c797 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Received event network-changed-f9078c39-0422-4f10-8f5e-12653d99d0ac {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.929413] env[61906]: DEBUG nova.compute.manager [req-f4be150e-5729-4bfe-9043-564407421cbc req-977b3865-5c22-4d9a-93d7-91027e61c797 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Refreshing instance network info cache due to event network-changed-f9078c39-0422-4f10-8f5e-12653d99d0ac. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 943.929708] env[61906]: DEBUG oslo_concurrency.lockutils [req-f4be150e-5729-4bfe-9043-564407421cbc req-977b3865-5c22-4d9a-93d7-91027e61c797 service nova] Acquiring lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.929957] env[61906]: DEBUG oslo_concurrency.lockutils [req-f4be150e-5729-4bfe-9043-564407421cbc req-977b3865-5c22-4d9a-93d7-91027e61c797 service nova] Acquired lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.930248] env[61906]: DEBUG nova.network.neutron [req-f4be150e-5729-4bfe-9043-564407421cbc req-977b3865-5c22-4d9a-93d7-91027e61c797 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Refreshing network info cache for port f9078c39-0422-4f10-8f5e-12653d99d0ac {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 943.996639] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.996932] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquired lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.997731] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 943.997731] env[61906]: DEBUG nova.objects.instance [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lazy-loading 'info_cache' on Instance uuid 79452791-59cb-4722-bb4a-8e59d8c4e641 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.102063] env[61906]: WARNING nova.compute.manager [None req-59a381fb-e5e4-48b2-8865-4126bbe52873 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Image not found during snapshot: nova.exception.ImageNotFound: Image 9a44efd7-e306-40d5-a03d-3f4b92997aba could not be found. [ 944.127217] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356936, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063823} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.127507] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 944.128350] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c00acfb-cd06-47f3-b5c1-e37a4e1a0948 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.148014] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] c77823f1-189c-4ac9-9a40-7bc45e82b442/c77823f1-189c-4ac9-9a40-7bc45e82b442.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 944.148832] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a7d4e73-9afe-45a5-9b01-3d5161ff8346 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.168389] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 944.168389] env[61906]: value = "task-1356938" [ 944.168389] env[61906]: _type = "Task" [ 944.168389] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.176611] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356938, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.190840] env[61906]: DEBUG oslo_vmware.api [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356937, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.582922] env[61906]: DEBUG oslo_concurrency.lockutils [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "070f8882-9099-4f17-bae8-1701b5dcc08d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.583216] env[61906]: DEBUG oslo_concurrency.lockutils [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "070f8882-9099-4f17-bae8-1701b5dcc08d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.583409] env[61906]: DEBUG oslo_concurrency.lockutils [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "070f8882-9099-4f17-bae8-1701b5dcc08d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.583597] env[61906]: DEBUG oslo_concurrency.lockutils [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "070f8882-9099-4f17-bae8-1701b5dcc08d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.583770] env[61906]: DEBUG oslo_concurrency.lockutils [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "070f8882-9099-4f17-bae8-1701b5dcc08d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.585848] env[61906]: INFO nova.compute.manager [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Terminating instance [ 944.587797] env[61906]: DEBUG nova.compute.manager [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 944.588151] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.589021] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3284a735-a4e3-4f11-84d7-df21a91ce29d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.597249] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.597498] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f647e675-8f58-4bd7-8585-9d900275698a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.603592] env[61906]: DEBUG oslo_vmware.api [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 944.603592] env[61906]: value = "task-1356939" [ 944.603592] env[61906]: _type = "Task" [ 944.603592] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.614308] env[61906]: DEBUG oslo_vmware.api [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356939, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.642815] env[61906]: DEBUG nova.network.neutron [req-f4be150e-5729-4bfe-9043-564407421cbc req-977b3865-5c22-4d9a-93d7-91027e61c797 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updated VIF entry in instance network info cache for port f9078c39-0422-4f10-8f5e-12653d99d0ac. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 944.643278] env[61906]: DEBUG nova.network.neutron [req-f4be150e-5729-4bfe-9043-564407421cbc req-977b3865-5c22-4d9a-93d7-91027e61c797 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updating instance_info_cache with network_info: [{"id": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "address": "fa:16:3e:b8:da:a6", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a08ec03-40", "ovs_interfaceid": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f9078c39-0422-4f10-8f5e-12653d99d0ac", "address": "fa:16:3e:57:10:a2", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9078c39-04", "ovs_interfaceid": "f9078c39-0422-4f10-8f5e-12653d99d0ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.678218] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356938, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.688544] env[61906]: DEBUG oslo_vmware.api [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356937, 'name': ReconfigVM_Task, 'duration_secs': 0.757434} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.689071] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.689301] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Reconfigured VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 944.691838] env[61906]: DEBUG oslo_concurrency.lockutils [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "766e79cb-bea3-4b74-923d-a147c7be4134" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.692084] env[61906]: DEBUG oslo_concurrency.lockutils [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "766e79cb-bea3-4b74-923d-a147c7be4134" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.692302] env[61906]: DEBUG oslo_concurrency.lockutils [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "766e79cb-bea3-4b74-923d-a147c7be4134-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.692493] env[61906]: DEBUG oslo_concurrency.lockutils [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "766e79cb-bea3-4b74-923d-a147c7be4134-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.692663] env[61906]: DEBUG oslo_concurrency.lockutils [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "766e79cb-bea3-4b74-923d-a147c7be4134-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.694440] env[61906]: INFO nova.compute.manager [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Terminating instance [ 944.696200] env[61906]: DEBUG nova.compute.manager [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 944.696396] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.697199] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe02d88-7699-4b92-a59f-a7a678c2de00 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.704170] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 944.704447] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f0c91706-27ff-485d-ba46-ef471df57620 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.768910] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 944.769204] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 944.769399] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleting the datastore file [datastore2] 766e79cb-bea3-4b74-923d-a147c7be4134 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 944.769668] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c477665-b71b-4dc7-94dd-0d1948d09e8a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.776720] env[61906]: DEBUG oslo_vmware.api [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 944.776720] env[61906]: value = "task-1356941" [ 944.776720] env[61906]: _type = "Task" [ 944.776720] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.784849] env[61906]: DEBUG oslo_vmware.api [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356941, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.113551] env[61906]: DEBUG oslo_vmware.api [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356939, 'name': PowerOffVM_Task, 'duration_secs': 0.222928} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.113845] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.114039] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.114276] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-85be8047-cec1-4331-af53-45a3c599f513 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.145943] env[61906]: DEBUG oslo_concurrency.lockutils [req-f4be150e-5729-4bfe-9043-564407421cbc req-977b3865-5c22-4d9a-93d7-91027e61c797 service nova] Releasing lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.179186] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356938, 'name': ReconfigVM_Task, 'duration_secs': 0.56948} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.180408] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Reconfigured VM instance instance-0000005b to attach disk [datastore2] c77823f1-189c-4ac9-9a40-7bc45e82b442/c77823f1-189c-4ac9-9a40-7bc45e82b442.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 945.181061] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.181269] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.181483] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleting the datastore file [datastore1] 070f8882-9099-4f17-bae8-1701b5dcc08d {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.181723] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-983b3431-7e1f-4eb7-ab9c-6477572aeefa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.183328] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e57ffbf-985e-4f70-9697-03683a2071b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.189464] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 945.189464] env[61906]: value = "task-1356943" [ 945.189464] env[61906]: _type = "Task" [ 945.189464] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.190641] env[61906]: DEBUG oslo_vmware.api [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for the task: (returnval){ [ 945.190641] env[61906]: value = "task-1356944" [ 945.190641] env[61906]: _type = "Task" [ 945.190641] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.197776] env[61906]: DEBUG oslo_concurrency.lockutils [None req-57093d1a-3734-466c-b9da-242ed83e69d5 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-99e5c55d-79ae-47a6-8500-79fc68291650-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.370s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.201945] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356943, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.205732] env[61906]: DEBUG oslo_vmware.api [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356944, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.287166] env[61906]: DEBUG oslo_vmware.api [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356941, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185529} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.287410] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 945.287587] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 945.287772] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.287984] env[61906]: INFO nova.compute.manager [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Took 0.59 seconds to destroy the instance on the hypervisor. [ 945.288262] env[61906]: DEBUG oslo.service.loopingcall [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.288468] env[61906]: DEBUG nova.compute.manager [-] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 945.288580] env[61906]: DEBUG nova.network.neutron [-] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 945.391105] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Volume attach. Driver type: vmdk {{(pid=61906) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 945.391794] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-289027', 'volume_id': '04f4539c-399e-4b60-be5b-43e8125b1e8a', 'name': 'volume-04f4539c-399e-4b60-be5b-43e8125b1e8a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd060a9f7-bd96-4e95-8780-1617a6ca7443', 'attached_at': '', 'detached_at': '', 'volume_id': '04f4539c-399e-4b60-be5b-43e8125b1e8a', 'serial': '04f4539c-399e-4b60-be5b-43e8125b1e8a'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 945.392262] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a718926b-1019-4e15-9aab-c901273e00a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.408987] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff6bad1-b820-44c6-8d93-4945d4d32b6e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.433591] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] volume-04f4539c-399e-4b60-be5b-43e8125b1e8a/volume-04f4539c-399e-4b60-be5b-43e8125b1e8a.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.433919] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6440a511-cb1c-4599-891b-56000f45988d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.451972] env[61906]: DEBUG oslo_vmware.api [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 945.451972] env[61906]: value = "task-1356945" [ 945.451972] env[61906]: _type = "Task" [ 945.451972] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.459995] env[61906]: DEBUG oslo_vmware.api [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356945, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.700233] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356943, 'name': Rename_Task, 'duration_secs': 0.186438} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.700837] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 945.701100] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b56b435-45fb-4583-9df7-68acf8d2172f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.706404] env[61906]: DEBUG oslo_vmware.api [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Task: {'id': task-1356944, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.308309} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.707477] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 945.707477] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 945.707477] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.707712] env[61906]: INFO nova.compute.manager [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 945.707900] env[61906]: DEBUG oslo.service.loopingcall [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.708279] env[61906]: DEBUG nova.compute.manager [-] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 945.708380] env[61906]: DEBUG nova.network.neutron [-] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 945.711389] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 945.711389] env[61906]: value = "task-1356946" [ 945.711389] env[61906]: _type = "Task" [ 945.711389] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.720083] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356946, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.759263] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Updating instance_info_cache with network_info: [{"id": "e291dedf-5c3f-4d78-bad3-295b2d1e2f75", "address": "fa:16:3e:c8:8d:ef", "network": {"id": "acdb1723-c152-4fe7-943b-096259fae6a6", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1911664852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24249dfdffc046ddb41d0da91203fa4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape291dedf-5c", "ovs_interfaceid": "e291dedf-5c3f-4d78-bad3-295b2d1e2f75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.961783] env[61906]: DEBUG nova.compute.manager [req-fdff5d28-e6a5-436f-974a-9d8fb06584db req-35ccc4e3-69ce-48dd-88d1-27337864342f service nova] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Received event network-vif-deleted-ce54463c-a7f7-4c89-8d23-cf62086bfaa3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.962049] env[61906]: INFO nova.compute.manager [req-fdff5d28-e6a5-436f-974a-9d8fb06584db req-35ccc4e3-69ce-48dd-88d1-27337864342f service nova] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Neutron deleted interface ce54463c-a7f7-4c89-8d23-cf62086bfaa3; detaching it from the instance and deleting it from the info cache [ 945.962409] env[61906]: DEBUG nova.network.neutron [req-fdff5d28-e6a5-436f-974a-9d8fb06584db req-35ccc4e3-69ce-48dd-88d1-27337864342f service nova] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.969429] env[61906]: DEBUG oslo_vmware.api [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356945, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.067588] env[61906]: DEBUG nova.network.neutron [-] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.229675] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356946, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.262456] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Releasing lock "refresh_cache-79452791-59cb-4722-bb4a-8e59d8c4e641" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.262750] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 946.262933] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.263129] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.263239] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.263398] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.263535] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.263696] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.263830] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 946.263977] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.462528] env[61906]: DEBUG oslo_vmware.api [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356945, 'name': ReconfigVM_Task, 'duration_secs': 0.588031} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.462828] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Reconfigured VM instance instance-00000056 to attach disk [datastore2] volume-04f4539c-399e-4b60-be5b-43e8125b1e8a/volume-04f4539c-399e-4b60-be5b-43e8125b1e8a.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.467693] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d362eb3f-383b-4d84-8ac9-defb9104fb5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.477486] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2fae3380-25dc-49a4-b419-404ae6a4565e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.483988] env[61906]: DEBUG oslo_vmware.api [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 946.483988] env[61906]: value = "task-1356947" [ 946.483988] env[61906]: _type = "Task" [ 946.483988] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.490520] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4ab7c0-c20e-474c-a341-62b3eeb5661c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.507482] env[61906]: DEBUG nova.network.neutron [-] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.508911] env[61906]: DEBUG oslo_vmware.api [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356947, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.520989] env[61906]: DEBUG nova.compute.manager [req-fdff5d28-e6a5-436f-974a-9d8fb06584db req-35ccc4e3-69ce-48dd-88d1-27337864342f service nova] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Detach interface failed, port_id=ce54463c-a7f7-4c89-8d23-cf62086bfaa3, reason: Instance 766e79cb-bea3-4b74-923d-a147c7be4134 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 946.569991] env[61906]: INFO nova.compute.manager [-] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Took 1.28 seconds to deallocate network for instance. [ 946.721423] env[61906]: DEBUG oslo_vmware.api [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356946, 'name': PowerOnVM_Task, 'duration_secs': 0.588775} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.721648] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 946.721848] env[61906]: DEBUG nova.compute.manager [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 946.722619] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22123ee5-7b60-4c95-b7c9-437fdc77828e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.768155] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.768373] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.768537] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.768685] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 946.769572] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99970355-3a1c-426d-8d1c-493983a940a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.777681] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc78c460-118c-4eb4-82e3-633dd2629105 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.791215] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73db75f-d4e8-4278-b7ce-ee268ecf4b92 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.797766] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a788b2b-5b82-41d2-9930-244b12144a0c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.825319] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180419MB free_disk=150GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 946.825459] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.825640] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.887727] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "interface-99e5c55d-79ae-47a6-8500-79fc68291650-7a45e277-aa49-45d3-a2d3-60792db87714" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.888048] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-99e5c55d-79ae-47a6-8500-79fc68291650-7a45e277-aa49-45d3-a2d3-60792db87714" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.888436] env[61906]: DEBUG nova.objects.instance [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'flavor' on Instance uuid 99e5c55d-79ae-47a6-8500-79fc68291650 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.993090] env[61906]: DEBUG oslo_vmware.api [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356947, 'name': ReconfigVM_Task, 'duration_secs': 0.136639} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.993429] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-289027', 'volume_id': '04f4539c-399e-4b60-be5b-43e8125b1e8a', 'name': 'volume-04f4539c-399e-4b60-be5b-43e8125b1e8a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd060a9f7-bd96-4e95-8780-1617a6ca7443', 'attached_at': '', 'detached_at': '', 'volume_id': '04f4539c-399e-4b60-be5b-43e8125b1e8a', 'serial': '04f4539c-399e-4b60-be5b-43e8125b1e8a'} {{(pid=61906) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 947.009650] env[61906]: INFO nova.compute.manager [-] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Took 1.30 seconds to deallocate network for instance. [ 947.076807] env[61906]: DEBUG oslo_concurrency.lockutils [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.238059] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.499873] env[61906]: DEBUG nova.objects.instance [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'pci_requests' on Instance uuid 99e5c55d-79ae-47a6-8500-79fc68291650 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.516019] env[61906]: DEBUG oslo_concurrency.lockutils [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.856088] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 79452791-59cb-4722-bb4a-8e59d8c4e641 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 947.856263] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 7bae658e-b050-4639-b34e-c2671ef5c773 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 947.856388] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance e133605d-c630-4b9e-a314-bf496c853710 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 947.856507] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 947.856623] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 99e5c55d-79ae-47a6-8500-79fc68291650 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 947.856737] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance d060a9f7-bd96-4e95-8780-1617a6ca7443 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 947.856848] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance cb879dc0-3af7-4279-aa28-66e2b8b4286d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 947.856960] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 070f8882-9099-4f17-bae8-1701b5dcc08d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 947.857082] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance c77823f1-189c-4ac9-9a40-7bc45e82b442 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 947.857222] env[61906]: WARNING nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 766e79cb-bea3-4b74-923d-a147c7be4134 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 947.857410] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 947.857546] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 947.904392] env[61906]: DEBUG oslo_concurrency.lockutils [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquiring lock "c77823f1-189c-4ac9-9a40-7bc45e82b442" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.904649] env[61906]: DEBUG oslo_concurrency.lockutils [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Lock "c77823f1-189c-4ac9-9a40-7bc45e82b442" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.904863] env[61906]: DEBUG oslo_concurrency.lockutils [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquiring lock "c77823f1-189c-4ac9-9a40-7bc45e82b442-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.905084] env[61906]: DEBUG oslo_concurrency.lockutils [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Lock "c77823f1-189c-4ac9-9a40-7bc45e82b442-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.905268] env[61906]: DEBUG oslo_concurrency.lockutils [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Lock "c77823f1-189c-4ac9-9a40-7bc45e82b442-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.909203] env[61906]: INFO nova.compute.manager [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Terminating instance [ 947.910977] env[61906]: DEBUG oslo_concurrency.lockutils [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquiring lock "refresh_cache-c77823f1-189c-4ac9-9a40-7bc45e82b442" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.911159] env[61906]: DEBUG oslo_concurrency.lockutils [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquired lock "refresh_cache-c77823f1-189c-4ac9-9a40-7bc45e82b442" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.911327] env[61906]: DEBUG nova.network.neutron [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 947.975102] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc393ba3-9ac4-472b-a7ea-51b5c36d04ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.981529] env[61906]: DEBUG nova.compute.manager [req-024a60b1-2a45-4c12-bda1-a8efc0038b6d req-fc857feb-3b63-4e4d-a307-4945c57d27e5 service nova] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Received event network-vif-deleted-e650d2e9-acd3-479b-9d48-748636ace077 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.984625] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06db903e-f367-4de8-a2ed-c968834e16ab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.018889] env[61906]: DEBUG nova.objects.base [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Object Instance<99e5c55d-79ae-47a6-8500-79fc68291650> lazy-loaded attributes: flavor,pci_requests {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 948.019146] env[61906]: DEBUG nova.network.neutron [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 948.021867] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153d5416-9f3e-4727-860c-d60ff423c32e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.030048] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793392c7-0554-4994-90e6-7f23a4a9d440 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.034165] env[61906]: DEBUG nova.objects.instance [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lazy-loading 'flavor' on Instance uuid d060a9f7-bd96-4e95-8780-1617a6ca7443 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.045032] env[61906]: DEBUG nova.compute.provider_tree [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.243733] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.264424] env[61906]: DEBUG nova.policy [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73219f5e96f64d1da86a19da9b18b06a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5533be74874b4094b7ef0f98a121f60b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 948.434688] env[61906]: DEBUG nova.network.neutron [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 948.511455] env[61906]: DEBUG nova.network.neutron [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.540212] env[61906]: DEBUG oslo_concurrency.lockutils [None req-47a02ee1-84a1-44bb-9dd2-e251c5f8686e tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.751s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.540212] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.297s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.550987] env[61906]: DEBUG nova.scheduler.client.report [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.014067] env[61906]: DEBUG oslo_concurrency.lockutils [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Releasing lock "refresh_cache-c77823f1-189c-4ac9-9a40-7bc45e82b442" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.014067] env[61906]: DEBUG nova.compute.manager [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 949.014067] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 949.015064] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-070a8c9b-1d94-409f-8b2c-d74bee236f88 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.022869] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 949.023128] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf7b07c1-d22f-455d-9650-0367ca9dfcb3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.029251] env[61906]: DEBUG oslo_vmware.api [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 949.029251] env[61906]: value = "task-1356948" [ 949.029251] env[61906]: _type = "Task" [ 949.029251] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.036755] env[61906]: DEBUG oslo_vmware.api [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356948, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.045322] env[61906]: INFO nova.compute.manager [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Detaching volume 04f4539c-399e-4b60-be5b-43e8125b1e8a [ 949.053438] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 949.053644] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.228s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.054147] env[61906]: DEBUG oslo_concurrency.lockutils [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.978s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.054364] env[61906]: DEBUG oslo_concurrency.lockutils [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.056196] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.818s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.056387] env[61906]: DEBUG nova.objects.instance [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61906) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 949.075451] env[61906]: INFO nova.scheduler.client.report [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleted allocations for instance 766e79cb-bea3-4b74-923d-a147c7be4134 [ 949.086012] env[61906]: INFO nova.virt.block_device [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Attempting to driver detach volume 04f4539c-399e-4b60-be5b-43e8125b1e8a from mountpoint /dev/sdb [ 949.086104] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 949.086276] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-289027', 'volume_id': '04f4539c-399e-4b60-be5b-43e8125b1e8a', 'name': 'volume-04f4539c-399e-4b60-be5b-43e8125b1e8a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd060a9f7-bd96-4e95-8780-1617a6ca7443', 'attached_at': '', 'detached_at': '', 'volume_id': '04f4539c-399e-4b60-be5b-43e8125b1e8a', 'serial': '04f4539c-399e-4b60-be5b-43e8125b1e8a'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 949.087199] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350838a5-8e1d-4b6a-a53a-31efd9a97ba4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.108845] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5555f2aa-4c05-4a9f-b517-48a5647233aa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.116607] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef41a9f-ae4e-4f27-bbee-64949ec28aae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.138550] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1c4457-8eae-415a-83c2-f9fb6db8b20f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.153220] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] The volume has not been displaced from its original location: [datastore2] volume-04f4539c-399e-4b60-be5b-43e8125b1e8a/volume-04f4539c-399e-4b60-be5b-43e8125b1e8a.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 949.158308] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Reconfiguring VM instance instance-00000056 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 949.158584] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91fc5cc0-0ac0-4080-ba18-c2eba31c43fe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.176461] env[61906]: DEBUG oslo_vmware.api [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 949.176461] env[61906]: value = "task-1356949" [ 949.176461] env[61906]: _type = "Task" [ 949.176461] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.184380] env[61906]: DEBUG oslo_vmware.api [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356949, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.538844] env[61906]: DEBUG oslo_vmware.api [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356948, 'name': PowerOffVM_Task, 'duration_secs': 0.185174} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.539264] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 949.539264] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 949.539478] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6bde28dd-1ebf-454e-ac92-439418eb02a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.564319] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 949.564541] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 949.564729] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Deleting the datastore file [datastore2] c77823f1-189c-4ac9-9a40-7bc45e82b442 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.565204] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b6f1a0dc-c7ae-48fd-8277-ae8185a2b898 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.571419] env[61906]: DEBUG oslo_vmware.api [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for the task: (returnval){ [ 949.571419] env[61906]: value = "task-1356951" [ 949.571419] env[61906]: _type = "Task" [ 949.571419] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.582069] env[61906]: DEBUG oslo_vmware.api [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356951, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.582495] env[61906]: DEBUG oslo_concurrency.lockutils [None req-45a167be-216b-408a-9c09-c263d7df4623 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "766e79cb-bea3-4b74-923d-a147c7be4134" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.890s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.686230] env[61906]: DEBUG oslo_vmware.api [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356949, 'name': ReconfigVM_Task, 'duration_secs': 0.211655} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.686546] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Reconfigured VM instance instance-00000056 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 949.691132] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c6d6c66-03a9-4a7b-8065-88d04aa9dd32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.705664] env[61906]: DEBUG oslo_vmware.api [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 949.705664] env[61906]: value = "task-1356952" [ 949.705664] env[61906]: _type = "Task" [ 949.705664] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.713505] env[61906]: DEBUG oslo_vmware.api [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356952, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.874398] env[61906]: DEBUG nova.network.neutron [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Successfully updated port: 7a45e277-aa49-45d3-a2d3-60792db87714 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.067225] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9d6c761f-8a20-46e4-9706-07b89312b897 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.069115] env[61906]: DEBUG oslo_concurrency.lockutils [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.553s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.069356] env[61906]: DEBUG nova.objects.instance [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lazy-loading 'resources' on Instance uuid 070f8882-9099-4f17-bae8-1701b5dcc08d {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.082889] env[61906]: DEBUG oslo_vmware.api [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Task: {'id': task-1356951, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106758} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.083164] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.083357] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 950.083537] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 950.083712] env[61906]: INFO nova.compute.manager [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Took 1.07 seconds to destroy the instance on the hypervisor. [ 950.083954] env[61906]: DEBUG oslo.service.loopingcall [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.084649] env[61906]: DEBUG nova.compute.manager [-] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.084780] env[61906]: DEBUG nova.network.neutron [-] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 950.103141] env[61906]: DEBUG nova.network.neutron [-] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 950.111833] env[61906]: DEBUG nova.compute.manager [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Received event network-vif-plugged-7a45e277-aa49-45d3-a2d3-60792db87714 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.112033] env[61906]: DEBUG oslo_concurrency.lockutils [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] Acquiring lock "99e5c55d-79ae-47a6-8500-79fc68291650-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.112248] env[61906]: DEBUG oslo_concurrency.lockutils [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] Lock "99e5c55d-79ae-47a6-8500-79fc68291650-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.112458] env[61906]: DEBUG oslo_concurrency.lockutils [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] Lock "99e5c55d-79ae-47a6-8500-79fc68291650-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.112654] env[61906]: DEBUG nova.compute.manager [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] No waiting events found dispatching network-vif-plugged-7a45e277-aa49-45d3-a2d3-60792db87714 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 950.112829] env[61906]: WARNING nova.compute.manager [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Received unexpected event network-vif-plugged-7a45e277-aa49-45d3-a2d3-60792db87714 for instance with vm_state active and task_state None. [ 950.112993] env[61906]: DEBUG nova.compute.manager [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Received event network-changed-7a45e277-aa49-45d3-a2d3-60792db87714 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.113182] env[61906]: DEBUG nova.compute.manager [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Refreshing instance network info cache due to event network-changed-7a45e277-aa49-45d3-a2d3-60792db87714. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 950.113373] env[61906]: DEBUG oslo_concurrency.lockutils [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] Acquiring lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.113513] env[61906]: DEBUG oslo_concurrency.lockutils [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] Acquired lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.113670] env[61906]: DEBUG nova.network.neutron [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Refreshing network info cache for port 7a45e277-aa49-45d3-a2d3-60792db87714 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 950.217576] env[61906]: DEBUG oslo_vmware.api [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356952, 'name': ReconfigVM_Task, 'duration_secs': 0.230285} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.217924] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-289027', 'volume_id': '04f4539c-399e-4b60-be5b-43e8125b1e8a', 'name': 'volume-04f4539c-399e-4b60-be5b-43e8125b1e8a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd060a9f7-bd96-4e95-8780-1617a6ca7443', 'attached_at': '', 'detached_at': '', 'volume_id': '04f4539c-399e-4b60-be5b-43e8125b1e8a', 'serial': '04f4539c-399e-4b60-be5b-43e8125b1e8a'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 950.378621] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.605809] env[61906]: DEBUG nova.network.neutron [-] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.628713] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.628962] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.700384] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3738b5ad-b91a-4b4d-9978-2b5d424e69ac {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.710062] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ebba3f7-edde-45fc-ab49-c563bbcfd7e2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.740884] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e49d39f1-d705-4c24-b150-916c6c10f4cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.748591] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83606707-4c9f-4987-ac10-224ae7f67ce3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.762184] env[61906]: DEBUG nova.compute.provider_tree [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.766898] env[61906]: DEBUG nova.objects.instance [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lazy-loading 'flavor' on Instance uuid d060a9f7-bd96-4e95-8780-1617a6ca7443 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.855603] env[61906]: DEBUG nova.network.neutron [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Added VIF to instance network info cache for port 7a45e277-aa49-45d3-a2d3-60792db87714. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3546}} [ 950.856102] env[61906]: DEBUG nova.network.neutron [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updating instance_info_cache with network_info: [{"id": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "address": "fa:16:3e:b8:da:a6", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a08ec03-40", "ovs_interfaceid": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f9078c39-0422-4f10-8f5e-12653d99d0ac", "address": "fa:16:3e:57:10:a2", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9078c39-04", "ovs_interfaceid": "f9078c39-0422-4f10-8f5e-12653d99d0ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7a45e277-aa49-45d3-a2d3-60792db87714", "address": "fa:16:3e:43:26:19", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a45e277-aa", "ovs_interfaceid": "7a45e277-aa49-45d3-a2d3-60792db87714", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.108305] env[61906]: INFO nova.compute.manager [-] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Took 1.02 seconds to deallocate network for instance. [ 951.131980] env[61906]: DEBUG nova.compute.manager [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 951.265084] env[61906]: DEBUG nova.scheduler.client.report [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.358469] env[61906]: DEBUG oslo_concurrency.lockutils [req-9607fb6f-4ce8-42df-887f-16bf36d02b91 req-4644f62d-1928-4b6c-84fb-9d88ef4ee510 service nova] Releasing lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.358865] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.359078] env[61906]: DEBUG nova.network.neutron [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.617507] env[61906]: DEBUG oslo_concurrency.lockutils [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.651591] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.773484] env[61906]: DEBUG oslo_concurrency.lockutils [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.704s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.775645] env[61906]: DEBUG oslo_concurrency.lockutils [None req-804627e3-3e6b-4517-bcef-aca7f3bde7b9 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.236s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.777358] env[61906]: DEBUG oslo_concurrency.lockutils [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.160s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.777679] env[61906]: DEBUG nova.objects.instance [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Lazy-loading 'resources' on Instance uuid c77823f1-189c-4ac9-9a40-7bc45e82b442 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.792524] env[61906]: INFO nova.scheduler.client.report [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Deleted allocations for instance 070f8882-9099-4f17-bae8-1701b5dcc08d [ 951.894690] env[61906]: WARNING nova.network.neutron [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] b12c63e1-bd61-4da3-8d37-4a89206ba0f2 already exists in list: networks containing: ['b12c63e1-bd61-4da3-8d37-4a89206ba0f2']. ignoring it [ 951.894920] env[61906]: WARNING nova.network.neutron [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] b12c63e1-bd61-4da3-8d37-4a89206ba0f2 already exists in list: networks containing: ['b12c63e1-bd61-4da3-8d37-4a89206ba0f2']. ignoring it [ 951.895112] env[61906]: WARNING nova.network.neutron [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] b12c63e1-bd61-4da3-8d37-4a89206ba0f2 already exists in list: networks containing: ['b12c63e1-bd61-4da3-8d37-4a89206ba0f2']. ignoring it [ 951.895283] env[61906]: WARNING nova.network.neutron [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] 7a45e277-aa49-45d3-a2d3-60792db87714 already exists in list: port_ids containing: ['7a45e277-aa49-45d3-a2d3-60792db87714']. ignoring it [ 952.301336] env[61906]: DEBUG nova.network.neutron [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updating instance_info_cache with network_info: [{"id": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "address": "fa:16:3e:b8:da:a6", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a08ec03-40", "ovs_interfaceid": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f9078c39-0422-4f10-8f5e-12653d99d0ac", "address": "fa:16:3e:57:10:a2", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9078c39-04", "ovs_interfaceid": "f9078c39-0422-4f10-8f5e-12653d99d0ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7a45e277-aa49-45d3-a2d3-60792db87714", "address": "fa:16:3e:43:26:19", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a45e277-aa", "ovs_interfaceid": "7a45e277-aa49-45d3-a2d3-60792db87714", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.305074] env[61906]: DEBUG oslo_concurrency.lockutils [None req-22a1cae1-d4f3-4b22-a62e-b05894a6fd78 tempest-ImagesTestJSON-1764945609 tempest-ImagesTestJSON-1764945609-project-member] Lock "070f8882-9099-4f17-bae8-1701b5dcc08d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.721s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.402943] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-059cd230-cb14-4895-80ad-5b6b0acfcc7a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.411053] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cfbcfe5-0c8d-4315-aa21-80cf686a61cf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.440686] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e12960af-e02c-4395-8908-004276dc964e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.447914] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c5c070-ca31-4eb9-a224-43a50ea6b879 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.461431] env[61906]: DEBUG nova.compute.provider_tree [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.807458] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.808225] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.808409] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.809535] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf051c4-89e5-4243-9ca7-c42d92be709a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.828986] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.829284] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.829503] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "d060a9f7-bd96-4e95-8780-1617a6ca7443-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.829690] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.829861] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.831832] env[61906]: DEBUG nova.virt.hardware [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 952.832071] env[61906]: DEBUG nova.virt.hardware [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 952.832241] env[61906]: DEBUG nova.virt.hardware [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 952.832429] env[61906]: DEBUG nova.virt.hardware [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 952.832580] env[61906]: DEBUG nova.virt.hardware [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 952.832730] env[61906]: DEBUG nova.virt.hardware [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 952.832940] env[61906]: DEBUG nova.virt.hardware [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 952.833124] env[61906]: DEBUG nova.virt.hardware [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 952.833304] env[61906]: DEBUG nova.virt.hardware [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 952.833473] env[61906]: DEBUG nova.virt.hardware [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 952.833651] env[61906]: DEBUG nova.virt.hardware [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 952.839905] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Reconfiguring VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 952.840461] env[61906]: INFO nova.compute.manager [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Terminating instance [ 952.841903] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c97d750d-0558-4442-b4b8-e935ddbbd1fb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.853644] env[61906]: DEBUG nova.compute.manager [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 952.853828] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.854852] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70efcf51-952a-4ecd-90a6-3d6c679bf419 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.862278] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.862915] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54835443-d113-45a6-bbf9-eb43746ca73c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.866599] env[61906]: DEBUG oslo_vmware.api [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 952.866599] env[61906]: value = "task-1356953" [ 952.866599] env[61906]: _type = "Task" [ 952.866599] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.870582] env[61906]: DEBUG oslo_vmware.api [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 952.870582] env[61906]: value = "task-1356954" [ 952.870582] env[61906]: _type = "Task" [ 952.870582] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.876550] env[61906]: DEBUG oslo_vmware.api [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356953, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.881103] env[61906]: DEBUG oslo_vmware.api [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356954, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.964358] env[61906]: DEBUG nova.scheduler.client.report [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 953.379908] env[61906]: DEBUG oslo_vmware.api [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356953, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.383384] env[61906]: DEBUG oslo_vmware.api [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356954, 'name': PowerOffVM_Task, 'duration_secs': 0.178103} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.383643] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.383813] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.384200] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b107c113-97fb-4f10-b663-050d1fa17eba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.460083] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 953.460246] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 953.460456] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Deleting the datastore file [datastore1] d060a9f7-bd96-4e95-8780-1617a6ca7443 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.460698] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0273520b-d7db-430f-b3ab-04b9952c467a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.467399] env[61906]: DEBUG oslo_vmware.api [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for the task: (returnval){ [ 953.467399] env[61906]: value = "task-1356956" [ 953.467399] env[61906]: _type = "Task" [ 953.467399] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.471909] env[61906]: DEBUG oslo_concurrency.lockutils [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.694s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.473911] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.822s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.475357] env[61906]: INFO nova.compute.claims [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 953.484597] env[61906]: DEBUG oslo_vmware.api [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356956, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.499682] env[61906]: INFO nova.scheduler.client.report [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Deleted allocations for instance c77823f1-189c-4ac9-9a40-7bc45e82b442 [ 953.885786] env[61906]: DEBUG oslo_vmware.api [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356953, 'name': ReconfigVM_Task, 'duration_secs': 0.624145} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.887967] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.887967] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Reconfigured VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 953.981683] env[61906]: DEBUG oslo_vmware.api [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Task: {'id': task-1356956, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154519} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.982206] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.982409] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 953.982587] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 953.982762] env[61906]: INFO nova.compute.manager [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Took 1.13 seconds to destroy the instance on the hypervisor. [ 953.983009] env[61906]: DEBUG oslo.service.loopingcall [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.983344] env[61906]: DEBUG nova.compute.manager [-] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 953.983514] env[61906]: DEBUG nova.network.neutron [-] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 954.009784] env[61906]: DEBUG oslo_concurrency.lockutils [None req-476384e3-98ff-43cb-ae1d-1b083207e958 tempest-ServerShowV257Test-24967767 tempest-ServerShowV257Test-24967767-project-member] Lock "c77823f1-189c-4ac9-9a40-7bc45e82b442" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.105s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.394474] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2d5ff100-5772-4b43-a543-651cfac072e2 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-99e5c55d-79ae-47a6-8500-79fc68291650-7a45e277-aa49-45d3-a2d3-60792db87714" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.506s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.489530] env[61906]: DEBUG nova.compute.manager [req-e40e055b-7cb7-4b3b-8613-d8ad3522e379 req-4817f34a-31d7-421b-baab-85dab63c147e service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Received event network-vif-deleted-e4fb390e-8245-4c57-a3b3-ef4556435df6 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.489530] env[61906]: INFO nova.compute.manager [req-e40e055b-7cb7-4b3b-8613-d8ad3522e379 req-4817f34a-31d7-421b-baab-85dab63c147e service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Neutron deleted interface e4fb390e-8245-4c57-a3b3-ef4556435df6; detaching it from the instance and deleting it from the info cache [ 954.489530] env[61906]: DEBUG nova.network.neutron [req-e40e055b-7cb7-4b3b-8613-d8ad3522e379 req-4817f34a-31d7-421b-baab-85dab63c147e service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.613985] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23387dff-d26c-43cc-817a-5e1d4b3d1c0b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.623524] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eeb8fe3-4456-41ac-b698-5f2eb19272cc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.661542] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1577a848-b5f8-40df-9576-3d4885121c77 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.670975] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56deccf8-ef39-4e66-b8c9-29a49d44a1e2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.691625] env[61906]: DEBUG nova.compute.provider_tree [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.964634] env[61906]: DEBUG nova.network.neutron [-] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.992691] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a25a77c-9ea1-4999-afa7-f392906d61a9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.002689] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5872864a-c1d0-4b71-98b7-29812712b22d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.034618] env[61906]: DEBUG nova.compute.manager [req-e40e055b-7cb7-4b3b-8613-d8ad3522e379 req-4817f34a-31d7-421b-baab-85dab63c147e service nova] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Detach interface failed, port_id=e4fb390e-8245-4c57-a3b3-ef4556435df6, reason: Instance d060a9f7-bd96-4e95-8780-1617a6ca7443 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 955.200233] env[61906]: DEBUG nova.scheduler.client.report [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.468745] env[61906]: INFO nova.compute.manager [-] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Took 1.49 seconds to deallocate network for instance. [ 955.706546] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.232s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.707126] env[61906]: DEBUG nova.compute.manager [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 955.980042] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.980460] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.980831] env[61906]: DEBUG nova.objects.instance [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lazy-loading 'resources' on Instance uuid d060a9f7-bd96-4e95-8780-1617a6ca7443 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.212644] env[61906]: DEBUG nova.compute.utils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 956.214092] env[61906]: DEBUG nova.compute.manager [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 956.214282] env[61906]: DEBUG nova.network.neutron [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 956.239543] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "e3578f5a-bd05-41fa-92e8-61d0e4ceed3f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.239543] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e3578f5a-bd05-41fa-92e8-61d0e4ceed3f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.274855] env[61906]: DEBUG nova.policy [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf4fea47b929461785f4a487a56977a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3093466de87a45a2ac673c25bea19bc3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 956.371238] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "interface-99e5c55d-79ae-47a6-8500-79fc68291650-f9078c39-0422-4f10-8f5e-12653d99d0ac" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.371544] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-99e5c55d-79ae-47a6-8500-79fc68291650-f9078c39-0422-4f10-8f5e-12653d99d0ac" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.564119] env[61906]: DEBUG nova.network.neutron [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Successfully created port: 0c3f7a1f-6d39-4063-b093-29db43118645 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 956.657414] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ac4f93-e476-43c4-8d53-07bb47673907 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.665619] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad82f47-a9b6-4f66-8bd0-63ff7123b23e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.701175] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e2bea3-72ec-4380-a671-2cafca36bcb8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.710995] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb442c6-5ca7-4cfc-918d-077dacd50b95 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.717110] env[61906]: DEBUG nova.compute.manager [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 956.728409] env[61906]: DEBUG nova.compute.provider_tree [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.741516] env[61906]: DEBUG nova.compute.manager [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 956.877021] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.877021] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.877021] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a9324c-21fb-4a20-88f3-37ec29cade30 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.896651] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5510f3c-2929-4a5e-8ecd-b9dbb2672992 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.926537] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Reconfiguring VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 956.927205] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb66337b-e9d4-473f-9488-89521e0bcaf4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.947015] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 956.947015] env[61906]: value = "task-1356957" [ 956.947015] env[61906]: _type = "Task" [ 956.947015] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.957481] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356957, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.234325] env[61906]: DEBUG nova.scheduler.client.report [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.261824] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.458608] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.739681] env[61906]: DEBUG nova.compute.manager [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 957.742313] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.761s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.744060] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.482s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.745853] env[61906]: INFO nova.compute.claims [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 957.770928] env[61906]: DEBUG nova.virt.hardware [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 957.770928] env[61906]: DEBUG nova.virt.hardware [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 957.770928] env[61906]: DEBUG nova.virt.hardware [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 957.770928] env[61906]: DEBUG nova.virt.hardware [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 957.770928] env[61906]: DEBUG nova.virt.hardware [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 957.770928] env[61906]: DEBUG nova.virt.hardware [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 957.770928] env[61906]: DEBUG nova.virt.hardware [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 957.771269] env[61906]: DEBUG nova.virt.hardware [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 957.771269] env[61906]: DEBUG nova.virt.hardware [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 957.771539] env[61906]: DEBUG nova.virt.hardware [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 957.771583] env[61906]: DEBUG nova.virt.hardware [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 957.772920] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a6edf6-d076-4d18-82ca-48fb841277dd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.776365] env[61906]: INFO nova.scheduler.client.report [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Deleted allocations for instance d060a9f7-bd96-4e95-8780-1617a6ca7443 [ 957.783926] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa25e1f7-ed5a-4a4c-a04f-79cb9b594be8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.961224] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.100063] env[61906]: DEBUG nova.compute.manager [req-8d224168-794d-44ae-8012-7c756d138a15 req-5e74b1d3-2753-42d0-899d-44c7bb9702cd service nova] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Received event network-vif-plugged-0c3f7a1f-6d39-4063-b093-29db43118645 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.100338] env[61906]: DEBUG oslo_concurrency.lockutils [req-8d224168-794d-44ae-8012-7c756d138a15 req-5e74b1d3-2753-42d0-899d-44c7bb9702cd service nova] Acquiring lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.100580] env[61906]: DEBUG oslo_concurrency.lockutils [req-8d224168-794d-44ae-8012-7c756d138a15 req-5e74b1d3-2753-42d0-899d-44c7bb9702cd service nova] Lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.100753] env[61906]: DEBUG oslo_concurrency.lockutils [req-8d224168-794d-44ae-8012-7c756d138a15 req-5e74b1d3-2753-42d0-899d-44c7bb9702cd service nova] Lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.100927] env[61906]: DEBUG nova.compute.manager [req-8d224168-794d-44ae-8012-7c756d138a15 req-5e74b1d3-2753-42d0-899d-44c7bb9702cd service nova] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] No waiting events found dispatching network-vif-plugged-0c3f7a1f-6d39-4063-b093-29db43118645 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 958.101111] env[61906]: WARNING nova.compute.manager [req-8d224168-794d-44ae-8012-7c756d138a15 req-5e74b1d3-2753-42d0-899d-44c7bb9702cd service nova] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Received unexpected event network-vif-plugged-0c3f7a1f-6d39-4063-b093-29db43118645 for instance with vm_state building and task_state spawning. [ 958.191505] env[61906]: DEBUG nova.network.neutron [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Successfully updated port: 0c3f7a1f-6d39-4063-b093-29db43118645 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 958.285317] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c15efb54-dab6-4066-9329-45964c6756b0 tempest-AttachVolumeNegativeTest-1093016191 tempest-AttachVolumeNegativeTest-1093016191-project-member] Lock "d060a9f7-bd96-4e95-8780-1617a6ca7443" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.456s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.460898] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.496988] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Acquiring lock "d2f16b79-c3fa-4450-8cb4-e9537215c60c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.497326] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Lock "d2f16b79-c3fa-4450-8cb4-e9537215c60c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.698231] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.698376] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.698533] env[61906]: DEBUG nova.network.neutron [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 958.876248] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba877ad-13e6-4e18-92e0-717c7fecd4b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.884188] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86a8c58-e888-4aab-816e-b77d29422234 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.913453] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406eac4b-1d55-4ef2-9102-056ff1a364c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.920927] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18c00f4-807c-4d99-afb3-bdd730907694 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.935584] env[61906]: DEBUG nova.compute.provider_tree [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.958663] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.001048] env[61906]: DEBUG nova.compute.manager [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 959.229166] env[61906]: DEBUG nova.network.neutron [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 959.359801] env[61906]: DEBUG nova.network.neutron [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance_info_cache with network_info: [{"id": "0c3f7a1f-6d39-4063-b093-29db43118645", "address": "fa:16:3e:77:db:95", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c3f7a1f-6d", "ovs_interfaceid": "0c3f7a1f-6d39-4063-b093-29db43118645", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.438378] env[61906]: DEBUG nova.scheduler.client.report [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.461893] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.519401] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.863669] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.864020] env[61906]: DEBUG nova.compute.manager [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Instance network_info: |[{"id": "0c3f7a1f-6d39-4063-b093-29db43118645", "address": "fa:16:3e:77:db:95", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c3f7a1f-6d", "ovs_interfaceid": "0c3f7a1f-6d39-4063-b093-29db43118645", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 959.864467] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:db:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '30f1dacf-8988-41b8-aa8f-e9530f65ef46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c3f7a1f-6d39-4063-b093-29db43118645', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 959.872037] env[61906]: DEBUG oslo.service.loopingcall [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.872037] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 959.872233] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-305c2ad6-17d5-403a-9eb9-f7405208c701 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.891765] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 959.891765] env[61906]: value = "task-1356959" [ 959.891765] env[61906]: _type = "Task" [ 959.891765] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.902195] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356959, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.944277] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.200s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.945176] env[61906]: DEBUG nova.compute.manager [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 959.947406] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.428s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.948905] env[61906]: INFO nova.compute.claims [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 959.959710] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.126887] env[61906]: DEBUG nova.compute.manager [req-76655497-6070-4ce7-8bd2-e121f4c850b9 req-2a730240-42dc-4a89-918d-5d188a01e5eb service nova] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Received event network-changed-0c3f7a1f-6d39-4063-b093-29db43118645 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.127170] env[61906]: DEBUG nova.compute.manager [req-76655497-6070-4ce7-8bd2-e121f4c850b9 req-2a730240-42dc-4a89-918d-5d188a01e5eb service nova] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Refreshing instance network info cache due to event network-changed-0c3f7a1f-6d39-4063-b093-29db43118645. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 960.127454] env[61906]: DEBUG oslo_concurrency.lockutils [req-76655497-6070-4ce7-8bd2-e121f4c850b9 req-2a730240-42dc-4a89-918d-5d188a01e5eb service nova] Acquiring lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.127632] env[61906]: DEBUG oslo_concurrency.lockutils [req-76655497-6070-4ce7-8bd2-e121f4c850b9 req-2a730240-42dc-4a89-918d-5d188a01e5eb service nova] Acquired lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.127825] env[61906]: DEBUG nova.network.neutron [req-76655497-6070-4ce7-8bd2-e121f4c850b9 req-2a730240-42dc-4a89-918d-5d188a01e5eb service nova] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Refreshing network info cache for port 0c3f7a1f-6d39-4063-b093-29db43118645 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 960.401713] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356959, 'name': CreateVM_Task, 'duration_secs': 0.28526} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.402014] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 960.402565] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.402733] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.403072] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 960.403324] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f0634b8-26cd-4bb6-a8fd-7c72b991e45c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.407710] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 960.407710] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527817a3-bd9f-38c7-2487-31465350b2e3" [ 960.407710] env[61906]: _type = "Task" [ 960.407710] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.416801] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527817a3-bd9f-38c7-2487-31465350b2e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.453260] env[61906]: DEBUG nova.compute.utils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 960.454855] env[61906]: DEBUG nova.compute.manager [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 960.455409] env[61906]: DEBUG nova.network.neutron [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 960.468773] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.505064] env[61906]: DEBUG nova.policy [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ba1658b4ade413b98834102c39a255c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c6414fa082f49c2a568a772a256a8e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 960.837075] env[61906]: DEBUG nova.network.neutron [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Successfully created port: f8b6d4d1-151c-45eb-b1c6-383bfaedef5b {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 960.923282] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527817a3-bd9f-38c7-2487-31465350b2e3, 'name': SearchDatastore_Task, 'duration_secs': 0.009316} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.927029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.927029] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 960.927029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.927029] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.927029] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.927029] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5646656a-b2d2-4d68-b6d5-31563094a5f7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.934283] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.934471] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 960.935257] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d20d82c-beb2-46d1-a593-96a114c90f83 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.942360] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 960.942360] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5257129c-3429-fce9-81a3-bf452a3d8394" [ 960.942360] env[61906]: _type = "Task" [ 960.942360] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.953087] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5257129c-3429-fce9-81a3-bf452a3d8394, 'name': SearchDatastore_Task, 'duration_secs': 0.008142} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.954244] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cbc4122-05e1-463a-8663-1c7a1f1516d6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.962859] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 960.962859] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528c1616-2fc7-c6b7-5250-b2b410d6c46b" [ 960.962859] env[61906]: _type = "Task" [ 960.962859] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.968801] env[61906]: DEBUG nova.compute.manager [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 960.971270] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.982356] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528c1616-2fc7-c6b7-5250-b2b410d6c46b, 'name': SearchDatastore_Task, 'duration_secs': 0.007697} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.982356] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.982356] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 082b1a00-dd07-4d37-860c-6a91d6d40eaf/082b1a00-dd07-4d37-860c-6a91d6d40eaf.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 960.982356] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c61b11d-7389-42e2-a4bd-c694e31a1f7d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.987066] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 960.987066] env[61906]: value = "task-1356960" [ 960.987066] env[61906]: _type = "Task" [ 960.987066] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.997907] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356960, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.121646] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98a65d4a-11f5-4df0-aa5d-330bca51072e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.129881] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b285e7a6-8745-49a6-82e6-1d6115f5b77a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.164512] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32937cff-ed19-4350-8cb7-e35b63f98286 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.172541] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db8c214-862e-4483-ab50-4cb9696f6da6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.186848] env[61906]: DEBUG nova.compute.provider_tree [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.243864] env[61906]: DEBUG nova.network.neutron [req-76655497-6070-4ce7-8bd2-e121f4c850b9 req-2a730240-42dc-4a89-918d-5d188a01e5eb service nova] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updated VIF entry in instance network info cache for port 0c3f7a1f-6d39-4063-b093-29db43118645. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 961.244317] env[61906]: DEBUG nova.network.neutron [req-76655497-6070-4ce7-8bd2-e121f4c850b9 req-2a730240-42dc-4a89-918d-5d188a01e5eb service nova] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance_info_cache with network_info: [{"id": "0c3f7a1f-6d39-4063-b093-29db43118645", "address": "fa:16:3e:77:db:95", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c3f7a1f-6d", "ovs_interfaceid": "0c3f7a1f-6d39-4063-b093-29db43118645", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.467442] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.501023] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356960, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464315} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.501023] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 082b1a00-dd07-4d37-860c-6a91d6d40eaf/082b1a00-dd07-4d37-860c-6a91d6d40eaf.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 961.503312] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.503640] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1f81983a-edad-4b2b-8508-e789824b71a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.511086] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 961.511086] env[61906]: value = "task-1356961" [ 961.511086] env[61906]: _type = "Task" [ 961.511086] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.520333] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356961, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.694883] env[61906]: DEBUG nova.scheduler.client.report [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 961.746903] env[61906]: DEBUG oslo_concurrency.lockutils [req-76655497-6070-4ce7-8bd2-e121f4c850b9 req-2a730240-42dc-4a89-918d-5d188a01e5eb service nova] Releasing lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.967212] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.983801] env[61906]: DEBUG nova.compute.manager [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 962.009755] env[61906]: DEBUG nova.virt.hardware [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.010023] env[61906]: DEBUG nova.virt.hardware [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.010621] env[61906]: DEBUG nova.virt.hardware [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.010621] env[61906]: DEBUG nova.virt.hardware [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.010621] env[61906]: DEBUG nova.virt.hardware [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.010803] env[61906]: DEBUG nova.virt.hardware [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.010891] env[61906]: DEBUG nova.virt.hardware [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.011076] env[61906]: DEBUG nova.virt.hardware [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.011283] env[61906]: DEBUG nova.virt.hardware [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.011486] env[61906]: DEBUG nova.virt.hardware [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.011698] env[61906]: DEBUG nova.virt.hardware [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.012611] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8596a389-f83e-4cf4-9996-cfb74dbd788f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.024911] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20c4886-1a6c-405f-b360-7a56eb84ec81 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.028661] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356961, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.053709} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.028921] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 962.030043] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa33781-a4bf-44fa-9db8-667770a4b3e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.059160] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 082b1a00-dd07-4d37-860c-6a91d6d40eaf/082b1a00-dd07-4d37-860c-6a91d6d40eaf.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.059454] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fec265ed-cc8b-4fe0-b53a-c8b1af3db92d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.077932] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 962.077932] env[61906]: value = "task-1356963" [ 962.077932] env[61906]: _type = "Task" [ 962.077932] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.085648] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356963, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.200424] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.467982] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356957, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.588148] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356963, 'name': ReconfigVM_Task, 'duration_secs': 0.353238} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.588545] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 082b1a00-dd07-4d37-860c-6a91d6d40eaf/082b1a00-dd07-4d37-860c-6a91d6d40eaf.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 962.589274] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c5f3e737-823d-4041-8bd6-dd1d08eeebe3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.595466] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 962.595466] env[61906]: value = "task-1356964" [ 962.595466] env[61906]: _type = "Task" [ 962.595466] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.603775] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356964, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.649829] env[61906]: DEBUG nova.compute.manager [req-476d4d78-d8ab-4da4-a8a1-64ddf32bd1ec req-5c0dc139-6fb3-4846-b6ce-4559b1657394 service nova] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Received event network-vif-plugged-f8b6d4d1-151c-45eb-b1c6-383bfaedef5b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.650029] env[61906]: DEBUG oslo_concurrency.lockutils [req-476d4d78-d8ab-4da4-a8a1-64ddf32bd1ec req-5c0dc139-6fb3-4846-b6ce-4559b1657394 service nova] Acquiring lock "e3578f5a-bd05-41fa-92e8-61d0e4ceed3f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.650251] env[61906]: DEBUG oslo_concurrency.lockutils [req-476d4d78-d8ab-4da4-a8a1-64ddf32bd1ec req-5c0dc139-6fb3-4846-b6ce-4559b1657394 service nova] Lock "e3578f5a-bd05-41fa-92e8-61d0e4ceed3f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.650429] env[61906]: DEBUG oslo_concurrency.lockutils [req-476d4d78-d8ab-4da4-a8a1-64ddf32bd1ec req-5c0dc139-6fb3-4846-b6ce-4559b1657394 service nova] Lock "e3578f5a-bd05-41fa-92e8-61d0e4ceed3f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.650599] env[61906]: DEBUG nova.compute.manager [req-476d4d78-d8ab-4da4-a8a1-64ddf32bd1ec req-5c0dc139-6fb3-4846-b6ce-4559b1657394 service nova] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] No waiting events found dispatching network-vif-plugged-f8b6d4d1-151c-45eb-b1c6-383bfaedef5b {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 962.650765] env[61906]: WARNING nova.compute.manager [req-476d4d78-d8ab-4da4-a8a1-64ddf32bd1ec req-5c0dc139-6fb3-4846-b6ce-4559b1657394 service nova] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Received unexpected event network-vif-plugged-f8b6d4d1-151c-45eb-b1c6-383bfaedef5b for instance with vm_state building and task_state spawning. [ 962.703603] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Acquiring lock "80fc23f4-8b8e-40c9-933a-3a00bcedd706" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.703864] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Lock "80fc23f4-8b8e-40c9-933a-3a00bcedd706" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.744807] env[61906]: DEBUG nova.network.neutron [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Successfully updated port: f8b6d4d1-151c-45eb-b1c6-383bfaedef5b {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 962.968019] env[61906]: DEBUG oslo_vmware.api [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356957, 'name': ReconfigVM_Task, 'duration_secs': 5.734237} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.968382] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.968691] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Reconfigured VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 963.106424] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356964, 'name': Rename_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.206509] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Lock "80fc23f4-8b8e-40c9-933a-3a00bcedd706" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.502s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.207088] env[61906]: DEBUG nova.compute.manager [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 963.246853] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-e3578f5a-bd05-41fa-92e8-61d0e4ceed3f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.247067] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-e3578f5a-bd05-41fa-92e8-61d0e4ceed3f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.247278] env[61906]: DEBUG nova.network.neutron [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 963.607612] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356964, 'name': Rename_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.637336] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.637603] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.711428] env[61906]: DEBUG nova.compute.utils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 963.712824] env[61906]: DEBUG nova.compute.manager [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 963.712996] env[61906]: DEBUG nova.network.neutron [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 963.818691] env[61906]: DEBUG nova.policy [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '035d3c289eed4312966351a9fb906034', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6fee48b37ad541409924509d9955a4fe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 963.827820] env[61906]: DEBUG nova.network.neutron [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 964.109858] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356964, 'name': Rename_Task, 'duration_secs': 1.156364} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.112946] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 964.115026] env[61906]: DEBUG nova.network.neutron [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Updating instance_info_cache with network_info: [{"id": "f8b6d4d1-151c-45eb-b1c6-383bfaedef5b", "address": "fa:16:3e:0c:6b:34", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8b6d4d1-15", "ovs_interfaceid": "f8b6d4d1-151c-45eb-b1c6-383bfaedef5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.115707] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab18cae0-e597-41ae-b6d6-23933c4fd2ba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.123182] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 964.123182] env[61906]: value = "task-1356965" [ 964.123182] env[61906]: _type = "Task" [ 964.123182] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.134898] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356965, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.140340] env[61906]: INFO nova.compute.manager [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Detaching volume cfc83076-1277-4f42-9c90-ef2e9aec2b8a [ 964.177025] env[61906]: INFO nova.virt.block_device [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Attempting to driver detach volume cfc83076-1277-4f42-9c90-ef2e9aec2b8a from mountpoint /dev/sdb [ 964.177106] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 964.177267] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-289013', 'volume_id': 'cfc83076-1277-4f42-9c90-ef2e9aec2b8a', 'name': 'volume-cfc83076-1277-4f42-9c90-ef2e9aec2b8a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63', 'attached_at': '', 'detached_at': '', 'volume_id': 'cfc83076-1277-4f42-9c90-ef2e9aec2b8a', 'serial': 'cfc83076-1277-4f42-9c90-ef2e9aec2b8a'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 964.178148] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b0cbcd-fc84-454c-b6d4-e5d7eeed6cbe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.181048] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.181226] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.181406] env[61906]: DEBUG nova.network.neutron [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 964.202259] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-268c8e07-717c-43b9-ae78-237d1f3e72ee {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.209900] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db7baf4-06bc-4f70-b5b4-f263549d8c19 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.237026] env[61906]: DEBUG nova.compute.manager [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 964.241142] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a10ae9-fa4e-4e1e-a47a-d3261014f413 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.260810] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] The volume has not been displaced from its original location: [datastore2] volume-cfc83076-1277-4f42-9c90-ef2e9aec2b8a/volume-cfc83076-1277-4f42-9c90-ef2e9aec2b8a.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 964.267753] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Reconfiguring VM instance instance-00000051 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 964.269088] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b109a448-160c-4db7-b5c8-8f6adb9ee86f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.292518] env[61906]: DEBUG oslo_vmware.api [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 964.292518] env[61906]: value = "task-1356966" [ 964.292518] env[61906]: _type = "Task" [ 964.292518] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.293747] env[61906]: DEBUG nova.network.neutron [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Successfully created port: 0ecf55ff-0f31-4c3b-9890-048efd2aa3a5 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 964.309334] env[61906]: DEBUG oslo_vmware.api [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356966, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.619211] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "99e5c55d-79ae-47a6-8500-79fc68291650" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.619560] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "99e5c55d-79ae-47a6-8500-79fc68291650" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.619853] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "99e5c55d-79ae-47a6-8500-79fc68291650-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.620120] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "99e5c55d-79ae-47a6-8500-79fc68291650-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.620377] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "99e5c55d-79ae-47a6-8500-79fc68291650-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.622300] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-e3578f5a-bd05-41fa-92e8-61d0e4ceed3f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.622638] env[61906]: DEBUG nova.compute.manager [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Instance network_info: |[{"id": "f8b6d4d1-151c-45eb-b1c6-383bfaedef5b", "address": "fa:16:3e:0c:6b:34", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8b6d4d1-15", "ovs_interfaceid": "f8b6d4d1-151c-45eb-b1c6-383bfaedef5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 964.623191] env[61906]: INFO nova.compute.manager [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Terminating instance [ 964.624616] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:6b:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8233b50c-be21-441a-a357-a29138a95b8b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f8b6d4d1-151c-45eb-b1c6-383bfaedef5b', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 964.631885] env[61906]: DEBUG oslo.service.loopingcall [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 964.632490] env[61906]: DEBUG nova.compute.manager [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 964.632682] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 964.632906] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 964.636865] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50bc1808-d9c0-40d1-bb04-47b752a88d06 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.639366] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-267dd8f8-8ece-4e14-9e3c-58ba57769208 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.658703] env[61906]: DEBUG oslo_vmware.api [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356965, 'name': PowerOnVM_Task, 'duration_secs': 0.484187} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.661714] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.661941] env[61906]: INFO nova.compute.manager [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Took 6.92 seconds to spawn the instance on the hypervisor. [ 964.662120] env[61906]: DEBUG nova.compute.manager [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.662358] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 964.662358] env[61906]: value = "task-1356967" [ 964.662358] env[61906]: _type = "Task" [ 964.662358] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.662586] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.663292] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5af6dfc-5aa2-4ee8-ba37-0cd3c77375ed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.665590] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc82e908-404e-4a8b-b52f-156ca872864f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.674540] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356967, 'name': CreateVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.679212] env[61906]: DEBUG oslo_vmware.api [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 964.679212] env[61906]: value = "task-1356968" [ 964.679212] env[61906]: _type = "Task" [ 964.679212] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.681787] env[61906]: DEBUG nova.compute.manager [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Received event network-changed-f8b6d4d1-151c-45eb-b1c6-383bfaedef5b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.681974] env[61906]: DEBUG nova.compute.manager [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Refreshing instance network info cache due to event network-changed-f8b6d4d1-151c-45eb-b1c6-383bfaedef5b. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 964.682204] env[61906]: DEBUG oslo_concurrency.lockutils [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] Acquiring lock "refresh_cache-e3578f5a-bd05-41fa-92e8-61d0e4ceed3f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.682350] env[61906]: DEBUG oslo_concurrency.lockutils [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] Acquired lock "refresh_cache-e3578f5a-bd05-41fa-92e8-61d0e4ceed3f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.682516] env[61906]: DEBUG nova.network.neutron [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Refreshing network info cache for port f8b6d4d1-151c-45eb-b1c6-383bfaedef5b {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 964.694143] env[61906]: DEBUG oslo_vmware.api [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356968, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.806475] env[61906]: DEBUG oslo_vmware.api [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356966, 'name': ReconfigVM_Task, 'duration_secs': 0.245592} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.809093] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Reconfigured VM instance instance-00000051 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 964.814256] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac3b7636-625e-42cd-bc5e-5cd9918ae7e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.830299] env[61906]: DEBUG oslo_vmware.api [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 964.830299] env[61906]: value = "task-1356969" [ 964.830299] env[61906]: _type = "Task" [ 964.830299] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.840459] env[61906]: DEBUG oslo_vmware.api [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356969, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.926317] env[61906]: INFO nova.network.neutron [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Port f9078c39-0422-4f10-8f5e-12653d99d0ac from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 964.926563] env[61906]: INFO nova.network.neutron [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Port 7a45e277-aa49-45d3-a2d3-60792db87714 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 964.926907] env[61906]: DEBUG nova.network.neutron [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updating instance_info_cache with network_info: [{"id": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "address": "fa:16:3e:b8:da:a6", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a08ec03-40", "ovs_interfaceid": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.175388] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356967, 'name': CreateVM_Task, 'duration_secs': 0.365247} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.175569] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 965.176223] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.176407] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.176740] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 965.177035] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ffda86e-1662-43ee-9170-7985b33e7648 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.185256] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 965.185256] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]520568ac-b837-e80e-d42b-d2804fe982b2" [ 965.185256] env[61906]: _type = "Task" [ 965.185256] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.195484] env[61906]: INFO nova.compute.manager [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Took 13.56 seconds to build instance. [ 965.202622] env[61906]: DEBUG oslo_vmware.api [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356968, 'name': PowerOffVM_Task, 'duration_secs': 0.205611} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.206885] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 965.206885] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 965.207968] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]520568ac-b837-e80e-d42b-d2804fe982b2, 'name': SearchDatastore_Task, 'duration_secs': 0.012128} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.207968] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-abe7d0dc-5184-4f51-baa8-2c3ff7a793d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.208959] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.209206] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 965.209449] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.209601] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.209791] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 965.210741] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-940bc5bd-d394-44f4-a3bf-a707e701ae9d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.219531] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 965.219714] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 965.220638] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d44afcae-bc95-4cbd-8363-a4a4c7cf5ed9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.230260] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 965.230260] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5203087f-e61e-7843-3dfa-242353fcf832" [ 965.230260] env[61906]: _type = "Task" [ 965.230260] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.238076] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5203087f-e61e-7843-3dfa-242353fcf832, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.247570] env[61906]: DEBUG nova.compute.manager [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 965.282241] env[61906]: DEBUG nova.virt.hardware [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 965.282518] env[61906]: DEBUG nova.virt.hardware [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 965.282681] env[61906]: DEBUG nova.virt.hardware [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.282870] env[61906]: DEBUG nova.virt.hardware [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 965.283038] env[61906]: DEBUG nova.virt.hardware [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.283201] env[61906]: DEBUG nova.virt.hardware [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 965.283559] env[61906]: DEBUG nova.virt.hardware [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 965.283882] env[61906]: DEBUG nova.virt.hardware [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 965.284185] env[61906]: DEBUG nova.virt.hardware [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 965.284464] env[61906]: DEBUG nova.virt.hardware [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 965.284813] env[61906]: DEBUG nova.virt.hardware [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.286152] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beeb07bb-b475-4be7-bc00-a59efca26705 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.298499] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd926885-565c-4bd3-ac5c-79fd8c1a810a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.339519] env[61906]: DEBUG oslo_vmware.api [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356969, 'name': ReconfigVM_Task, 'duration_secs': 0.147367} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.339879] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-289013', 'volume_id': 'cfc83076-1277-4f42-9c90-ef2e9aec2b8a', 'name': 'volume-cfc83076-1277-4f42-9c90-ef2e9aec2b8a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63', 'attached_at': '', 'detached_at': '', 'volume_id': 'cfc83076-1277-4f42-9c90-ef2e9aec2b8a', 'serial': 'cfc83076-1277-4f42-9c90-ef2e9aec2b8a'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 965.348044] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 965.348236] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 965.348450] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Deleting the datastore file [datastore2] 99e5c55d-79ae-47a6-8500-79fc68291650 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.348825] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66b0c168-19d2-41bd-b113-84eb4c52365f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.356391] env[61906]: DEBUG oslo_vmware.api [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 965.356391] env[61906]: value = "task-1356971" [ 965.356391] env[61906]: _type = "Task" [ 965.356391] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.367979] env[61906]: DEBUG oslo_vmware.api [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356971, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.428825] env[61906]: DEBUG nova.network.neutron [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Updated VIF entry in instance network info cache for port f8b6d4d1-151c-45eb-b1c6-383bfaedef5b. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 965.429318] env[61906]: DEBUG nova.network.neutron [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Updating instance_info_cache with network_info: [{"id": "f8b6d4d1-151c-45eb-b1c6-383bfaedef5b", "address": "fa:16:3e:0c:6b:34", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8b6d4d1-15", "ovs_interfaceid": "f8b6d4d1-151c-45eb-b1c6-383bfaedef5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.430973] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "refresh_cache-99e5c55d-79ae-47a6-8500-79fc68291650" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.698068] env[61906]: DEBUG oslo_concurrency.lockutils [None req-6c63c516-acc2-4cfc-984d-425bc45b24ef tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.069s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.741115] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5203087f-e61e-7843-3dfa-242353fcf832, 'name': SearchDatastore_Task, 'duration_secs': 0.012036} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.741932] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-642c7ea9-ffa7-400c-96b9-860709bc4638 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.748040] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 965.748040] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521bb3fa-60a9-4bb8-138f-f547e000abf4" [ 965.748040] env[61906]: _type = "Task" [ 965.748040] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.758875] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521bb3fa-60a9-4bb8-138f-f547e000abf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.866846] env[61906]: DEBUG oslo_vmware.api [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1356971, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.465294} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.867214] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.867337] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 965.867524] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 965.867705] env[61906]: INFO nova.compute.manager [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Took 1.24 seconds to destroy the instance on the hypervisor. [ 965.867954] env[61906]: DEBUG oslo.service.loopingcall [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.868166] env[61906]: DEBUG nova.compute.manager [-] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 965.868258] env[61906]: DEBUG nova.network.neutron [-] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 965.892211] env[61906]: DEBUG nova.objects.instance [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lazy-loading 'flavor' on Instance uuid 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.924893] env[61906]: DEBUG nova.network.neutron [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Successfully updated port: 0ecf55ff-0f31-4c3b-9890-048efd2aa3a5 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 965.935402] env[61906]: DEBUG oslo_concurrency.lockutils [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] Releasing lock "refresh_cache-e3578f5a-bd05-41fa-92e8-61d0e4ceed3f" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.935402] env[61906]: DEBUG nova.compute.manager [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Received event network-vif-deleted-f9078c39-0422-4f10-8f5e-12653d99d0ac {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.935402] env[61906]: INFO nova.compute.manager [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Neutron deleted interface f9078c39-0422-4f10-8f5e-12653d99d0ac; detaching it from the instance and deleting it from the info cache [ 965.935513] env[61906]: DEBUG nova.network.neutron [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updating instance_info_cache with network_info: [{"id": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "address": "fa:16:3e:b8:da:a6", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a08ec03-40", "ovs_interfaceid": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7a45e277-aa49-45d3-a2d3-60792db87714", "address": "fa:16:3e:43:26:19", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a45e277-aa", "ovs_interfaceid": "7a45e277-aa49-45d3-a2d3-60792db87714", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.937263] env[61906]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 7a45e277-aa49-45d3-a2d3-60792db87714 could not be found.", "detail": ""}} {{(pid=61906) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 965.937371] env[61906]: DEBUG nova.network.neutron [-] Unable to show port 7a45e277-aa49-45d3-a2d3-60792db87714 as it no longer exists. {{(pid=61906) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 965.946067] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7d1ae25d-c503-486d-bcf2-780af6042cf1 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-99e5c55d-79ae-47a6-8500-79fc68291650-f9078c39-0422-4f10-8f5e-12653d99d0ac" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.572s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.264261] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521bb3fa-60a9-4bb8-138f-f547e000abf4, 'name': SearchDatastore_Task, 'duration_secs': 0.016757} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.265084] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.265404] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] e3578f5a-bd05-41fa-92e8-61d0e4ceed3f/e3578f5a-bd05-41fa-92e8-61d0e4ceed3f.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 966.265726] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7bc154f7-612f-4e50-a7b5-60026c99e377 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.272916] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 966.272916] env[61906]: value = "task-1356972" [ 966.272916] env[61906]: _type = "Task" [ 966.272916] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.282225] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356972, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.426880] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Acquiring lock "refresh_cache-d2f16b79-c3fa-4450-8cb4-e9537215c60c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.431219] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Acquired lock "refresh_cache-d2f16b79-c3fa-4450-8cb4-e9537215c60c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.431446] env[61906]: DEBUG nova.network.neutron [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 966.449960] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-96c6b004-2f30-4d04-a7bc-5257f6eb52cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.462974] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2daba7b-d9a1-4b19-a6e3-11e2826c8c28 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.495120] env[61906]: DEBUG nova.compute.manager [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Detach interface failed, port_id=f9078c39-0422-4f10-8f5e-12653d99d0ac, reason: Instance 99e5c55d-79ae-47a6-8500-79fc68291650 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 966.495566] env[61906]: DEBUG nova.compute.manager [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Received event network-vif-deleted-7a45e277-aa49-45d3-a2d3-60792db87714 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.495789] env[61906]: INFO nova.compute.manager [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Neutron deleted interface 7a45e277-aa49-45d3-a2d3-60792db87714; detaching it from the instance and deleting it from the info cache [ 966.496068] env[61906]: DEBUG nova.network.neutron [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updating instance_info_cache with network_info: [{"id": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "address": "fa:16:3e:b8:da:a6", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a08ec03-40", "ovs_interfaceid": "9a08ec03-408f-4d57-842d-3f68f53dcb29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.717444] env[61906]: DEBUG nova.compute.manager [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Received event network-vif-plugged-0ecf55ff-0f31-4c3b-9890-048efd2aa3a5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.717744] env[61906]: DEBUG oslo_concurrency.lockutils [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] Acquiring lock "d2f16b79-c3fa-4450-8cb4-e9537215c60c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.717896] env[61906]: DEBUG oslo_concurrency.lockutils [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] Lock "d2f16b79-c3fa-4450-8cb4-e9537215c60c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.718094] env[61906]: DEBUG oslo_concurrency.lockutils [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] Lock "d2f16b79-c3fa-4450-8cb4-e9537215c60c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.718273] env[61906]: DEBUG nova.compute.manager [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] No waiting events found dispatching network-vif-plugged-0ecf55ff-0f31-4c3b-9890-048efd2aa3a5 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 966.718512] env[61906]: WARNING nova.compute.manager [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Received unexpected event network-vif-plugged-0ecf55ff-0f31-4c3b-9890-048efd2aa3a5 for instance with vm_state building and task_state spawning. [ 966.718695] env[61906]: DEBUG nova.compute.manager [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Received event network-changed-0ecf55ff-0f31-4c3b-9890-048efd2aa3a5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.718852] env[61906]: DEBUG nova.compute.manager [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Refreshing instance network info cache due to event network-changed-0ecf55ff-0f31-4c3b-9890-048efd2aa3a5. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 966.719364] env[61906]: DEBUG oslo_concurrency.lockutils [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] Acquiring lock "refresh_cache-d2f16b79-c3fa-4450-8cb4-e9537215c60c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.786600] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356972, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.875101] env[61906]: DEBUG nova.network.neutron [-] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.899560] env[61906]: DEBUG oslo_concurrency.lockutils [None req-9273eeee-d2af-45b0-bc8f-e3961fedc7ff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.262s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.968580] env[61906]: DEBUG nova.network.neutron [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 966.998593] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4673bacb-3254-41f0-9d92-e33bc6b71aa9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.007559] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f96667-bfb0-40e2-8129-ebe5061f83aa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.047344] env[61906]: DEBUG nova.compute.manager [req-3dba0c65-a605-4326-912b-645277cac27e req-64d5dc56-753b-4607-bafd-bc42d3c5434b service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Detach interface failed, port_id=7a45e277-aa49-45d3-a2d3-60792db87714, reason: Instance 99e5c55d-79ae-47a6-8500-79fc68291650 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 967.127928] env[61906]: DEBUG nova.network.neutron [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Updating instance_info_cache with network_info: [{"id": "0ecf55ff-0f31-4c3b-9890-048efd2aa3a5", "address": "fa:16:3e:4a:68:5c", "network": {"id": "0d0ef46f-e76e-4f8e-9e48-26dbe8ccd139", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-81806780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6fee48b37ad541409924509d9955a4fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ecf55ff-0f", "ovs_interfaceid": "0ecf55ff-0f31-4c3b-9890-048efd2aa3a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.284581] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356972, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.666563} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.284842] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] e3578f5a-bd05-41fa-92e8-61d0e4ceed3f/e3578f5a-bd05-41fa-92e8-61d0e4ceed3f.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 967.285097] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 967.285376] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cbce5930-8c3c-49a5-adba-1b33deab2b86 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.291545] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 967.291545] env[61906]: value = "task-1356973" [ 967.291545] env[61906]: _type = "Task" [ 967.291545] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.299855] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356973, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.315904] env[61906]: DEBUG nova.compute.manager [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Stashing vm_state: active {{(pid=61906) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 967.377975] env[61906]: INFO nova.compute.manager [-] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Took 1.51 seconds to deallocate network for instance. [ 967.631116] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Releasing lock "refresh_cache-d2f16b79-c3fa-4450-8cb4-e9537215c60c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.631463] env[61906]: DEBUG nova.compute.manager [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Instance network_info: |[{"id": "0ecf55ff-0f31-4c3b-9890-048efd2aa3a5", "address": "fa:16:3e:4a:68:5c", "network": {"id": "0d0ef46f-e76e-4f8e-9e48-26dbe8ccd139", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-81806780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6fee48b37ad541409924509d9955a4fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ecf55ff-0f", "ovs_interfaceid": "0ecf55ff-0f31-4c3b-9890-048efd2aa3a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 967.632095] env[61906]: DEBUG oslo_concurrency.lockutils [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] Acquired lock "refresh_cache-d2f16b79-c3fa-4450-8cb4-e9537215c60c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.632292] env[61906]: DEBUG nova.network.neutron [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Refreshing network info cache for port 0ecf55ff-0f31-4c3b-9890-048efd2aa3a5 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 967.633766] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:68:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '04e15990-16e1-4cb2-b0f0-06c362e68c5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0ecf55ff-0f31-4c3b-9890-048efd2aa3a5', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 967.641992] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Creating folder: Project (6fee48b37ad541409924509d9955a4fe). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 967.645080] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-65a67ea5-eed7-4d29-a260-000116a0afd5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.658619] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Created folder: Project (6fee48b37ad541409924509d9955a4fe) in parent group-v288914. [ 967.658830] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Creating folder: Instances. Parent ref: group-v289030. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 967.659054] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6f544fe7-d30c-4ccc-94ea-eaf2252b1185 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.667038] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Created folder: Instances in parent group-v289030. [ 967.667508] env[61906]: DEBUG oslo.service.loopingcall [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.667508] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 967.667664] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a707d99-1076-4a76-9409-7fc2dd023fcc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.687109] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 967.687109] env[61906]: value = "task-1356976" [ 967.687109] env[61906]: _type = "Task" [ 967.687109] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.694071] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356976, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.801848] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356973, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073999} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.802208] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 967.802917] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d370b4-d21f-4b5a-93ce-8f0734e52f2b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.823930] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] e3578f5a-bd05-41fa-92e8-61d0e4ceed3f/e3578f5a-bd05-41fa-92e8-61d0e4ceed3f.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 967.828880] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55e91848-d953-4499-8c35-e6335b27b670 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.848070] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 967.848070] env[61906]: value = "task-1356977" [ 967.848070] env[61906]: _type = "Task" [ 967.848070] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.856039] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356977, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.856942] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.857199] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.879716] env[61906]: DEBUG nova.network.neutron [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Updated VIF entry in instance network info cache for port 0ecf55ff-0f31-4c3b-9890-048efd2aa3a5. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 967.880133] env[61906]: DEBUG nova.network.neutron [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Updating instance_info_cache with network_info: [{"id": "0ecf55ff-0f31-4c3b-9890-048efd2aa3a5", "address": "fa:16:3e:4a:68:5c", "network": {"id": "0d0ef46f-e76e-4f8e-9e48-26dbe8ccd139", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-81806780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6fee48b37ad541409924509d9955a4fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "04e15990-16e1-4cb2-b0f0-06c362e68c5e", "external-id": "nsx-vlan-transportzone-555", "segmentation_id": 555, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ecf55ff-0f", "ovs_interfaceid": "0ecf55ff-0f31-4c3b-9890-048efd2aa3a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.883518] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.993956] env[61906]: DEBUG oslo_concurrency.lockutils [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.994163] env[61906]: DEBUG oslo_concurrency.lockutils [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.995058] env[61906]: DEBUG oslo_concurrency.lockutils [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.995058] env[61906]: DEBUG oslo_concurrency.lockutils [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.995058] env[61906]: DEBUG oslo_concurrency.lockutils [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.997794] env[61906]: INFO nova.compute.manager [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Terminating instance [ 967.999806] env[61906]: DEBUG nova.compute.manager [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.000057] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 968.001031] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b5ca6d5-cf6b-4b1a-b507-35dfc0b752be {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.012093] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.012352] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c373c97c-c183-46ce-b04a-4d7df845fd00 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.018184] env[61906]: DEBUG oslo_vmware.api [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 968.018184] env[61906]: value = "task-1356978" [ 968.018184] env[61906]: _type = "Task" [ 968.018184] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.026244] env[61906]: DEBUG oslo_vmware.api [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.198756] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356976, 'name': CreateVM_Task, 'duration_secs': 0.345851} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.199121] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 968.199897] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.200135] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.200476] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 968.200761] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a8c0887-92cd-40e1-9295-f51dba7e5e70 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.205745] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Waiting for the task: (returnval){ [ 968.205745] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]520323b1-29db-1c4b-3475-eedee3a4844d" [ 968.205745] env[61906]: _type = "Task" [ 968.205745] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.213267] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]520323b1-29db-1c4b-3475-eedee3a4844d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.357558] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356977, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.362276] env[61906]: INFO nova.compute.claims [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 968.382704] env[61906]: DEBUG oslo_concurrency.lockutils [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] Releasing lock "refresh_cache-d2f16b79-c3fa-4450-8cb4-e9537215c60c" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.383109] env[61906]: DEBUG nova.compute.manager [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Received event network-vif-deleted-9a08ec03-408f-4d57-842d-3f68f53dcb29 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.383320] env[61906]: INFO nova.compute.manager [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Neutron deleted interface 9a08ec03-408f-4d57-842d-3f68f53dcb29; detaching it from the instance and deleting it from the info cache [ 968.383498] env[61906]: DEBUG nova.network.neutron [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.529294] env[61906]: DEBUG oslo_vmware.api [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356978, 'name': PowerOffVM_Task, 'duration_secs': 0.193621} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.529604] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 968.529777] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 968.530057] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-718be019-7f74-4510-9369-25158dc569d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.678277] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 968.678589] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 968.678831] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Deleting the datastore file [datastore2] 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.679150] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-88d9f728-ba59-400a-8fff-0143b136067d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.685958] env[61906]: DEBUG oslo_vmware.api [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 968.685958] env[61906]: value = "task-1356980" [ 968.685958] env[61906]: _type = "Task" [ 968.685958] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.695152] env[61906]: DEBUG oslo_vmware.api [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356980, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.715744] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]520323b1-29db-1c4b-3475-eedee3a4844d, 'name': SearchDatastore_Task, 'duration_secs': 0.054402} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.716063] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.716403] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.716733] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.716969] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.717212] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.717514] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6cd08d52-2213-40c2-a758-9021bfe64dc2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.725085] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.725274] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 968.726161] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcad40a8-6b73-45e6-b948-8c8c299f9261 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.731421] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Waiting for the task: (returnval){ [ 968.731421] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5248d1dd-b4de-8d31-fae1-42fa5bb1b42c" [ 968.731421] env[61906]: _type = "Task" [ 968.731421] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.741413] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5248d1dd-b4de-8d31-fae1-42fa5bb1b42c, 'name': SearchDatastore_Task} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.742133] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-998b7d56-17d8-489d-b627-f00787c6df39 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.746649] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Waiting for the task: (returnval){ [ 968.746649] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52067c54-f5c4-2f74-0392-ca90cff432e7" [ 968.746649] env[61906]: _type = "Task" [ 968.746649] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.753480] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52067c54-f5c4-2f74-0392-ca90cff432e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.858292] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356977, 'name': ReconfigVM_Task, 'duration_secs': 0.904521} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.858648] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Reconfigured VM instance instance-0000005e to attach disk [datastore2] e3578f5a-bd05-41fa-92e8-61d0e4ceed3f/e3578f5a-bd05-41fa-92e8-61d0e4ceed3f.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 968.859289] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92ac9df2-a112-4d3e-8b39-60b148a6497f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.864661] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 968.864661] env[61906]: value = "task-1356981" [ 968.864661] env[61906]: _type = "Task" [ 968.864661] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.868789] env[61906]: INFO nova.compute.resource_tracker [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating resource usage from migration 64b6b34a-47b3-4cdf-b34b-951bf173cffb [ 968.875862] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356981, 'name': Rename_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.885870] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fb2d3584-293b-43d2-ac4d-a121df174d69 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.896228] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70c17cd-3308-497d-bfab-7066b06affce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.929399] env[61906]: DEBUG nova.compute.manager [req-def61cb5-d5cc-41f2-af41-0e649bbb2b54 req-cdeee9a3-424e-482d-8126-524fe0ba0fc4 service nova] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Detach interface failed, port_id=9a08ec03-408f-4d57-842d-3f68f53dcb29, reason: Instance 99e5c55d-79ae-47a6-8500-79fc68291650 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 969.005023] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ba3963-32fe-496e-9de7-51c74fc0239f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.011261] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0d3786-463d-48d8-a180-922f64691fe0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.040101] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14474ab0-9b75-4b68-b63a-ad680e48842e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.046969] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50c45f54-0b40-40c4-9276-b1537dc2ded5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.061410] env[61906]: DEBUG nova.compute.provider_tree [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.196388] env[61906]: DEBUG oslo_vmware.api [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1356980, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.291977} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.196704] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.196894] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 969.197089] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 969.197287] env[61906]: INFO nova.compute.manager [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Took 1.20 seconds to destroy the instance on the hypervisor. [ 969.197533] env[61906]: DEBUG oslo.service.loopingcall [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.197736] env[61906]: DEBUG nova.compute.manager [-] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 969.197829] env[61906]: DEBUG nova.network.neutron [-] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 969.258581] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52067c54-f5c4-2f74-0392-ca90cff432e7, 'name': SearchDatastore_Task, 'duration_secs': 0.008319} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.258855] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.259146] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] d2f16b79-c3fa-4450-8cb4-e9537215c60c/d2f16b79-c3fa-4450-8cb4-e9537215c60c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 969.259385] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb2a3ca0-0261-4a02-8630-5d6308f4b4e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.265890] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Waiting for the task: (returnval){ [ 969.265890] env[61906]: value = "task-1356982" [ 969.265890] env[61906]: _type = "Task" [ 969.265890] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.273519] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356982, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.373846] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356981, 'name': Rename_Task, 'duration_secs': 0.136725} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.374133] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.374380] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67ad2611-0a38-4249-8906-b06dcd821ce8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.379739] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 969.379739] env[61906]: value = "task-1356983" [ 969.379739] env[61906]: _type = "Task" [ 969.379739] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.387164] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356983, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.565359] env[61906]: DEBUG nova.scheduler.client.report [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.731729] env[61906]: DEBUG nova.compute.manager [req-ddeaacf4-386d-450f-8cdc-b02b8cb349a5 req-eaf3cec8-9969-4fc1-a25f-18d31a9fe741 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Received event network-vif-deleted-eb78ba42-034c-44d7-9dd3-4157cd702c8c {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.731729] env[61906]: INFO nova.compute.manager [req-ddeaacf4-386d-450f-8cdc-b02b8cb349a5 req-eaf3cec8-9969-4fc1-a25f-18d31a9fe741 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Neutron deleted interface eb78ba42-034c-44d7-9dd3-4157cd702c8c; detaching it from the instance and deleting it from the info cache [ 969.731729] env[61906]: DEBUG nova.network.neutron [req-ddeaacf4-386d-450f-8cdc-b02b8cb349a5 req-eaf3cec8-9969-4fc1-a25f-18d31a9fe741 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.777726] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356982, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.891021] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356983, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.072113] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.214s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.072113] env[61906]: INFO nova.compute.manager [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Migrating [ 970.078451] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.195s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.078697] env[61906]: DEBUG nova.objects.instance [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'resources' on Instance uuid 99e5c55d-79ae-47a6-8500-79fc68291650 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.208161] env[61906]: DEBUG nova.network.neutron [-] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.235234] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9561c81f-3738-4c70-814f-3df53279b8e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.247113] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4599b8-712d-4bc2-b6a0-2431f5e6a89c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.273729] env[61906]: DEBUG nova.compute.manager [req-ddeaacf4-386d-450f-8cdc-b02b8cb349a5 req-eaf3cec8-9969-4fc1-a25f-18d31a9fe741 service nova] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Detach interface failed, port_id=eb78ba42-034c-44d7-9dd3-4157cd702c8c, reason: Instance 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 970.284564] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356982, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.574967} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.285026] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] d2f16b79-c3fa-4450-8cb4-e9537215c60c/d2f16b79-c3fa-4450-8cb4-e9537215c60c.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 970.285365] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 970.285733] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-31eb4963-c5e2-43cc-851e-03774a1f4366 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.293363] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Waiting for the task: (returnval){ [ 970.293363] env[61906]: value = "task-1356984" [ 970.293363] env[61906]: _type = "Task" [ 970.293363] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.304042] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356984, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.389970] env[61906]: DEBUG oslo_vmware.api [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356983, 'name': PowerOnVM_Task, 'duration_secs': 0.697755} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.390320] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 970.390524] env[61906]: INFO nova.compute.manager [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Took 8.41 seconds to spawn the instance on the hypervisor. [ 970.390711] env[61906]: DEBUG nova.compute.manager [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.391492] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe28389-2f7c-445a-8aa0-ed412f3b3296 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.589554] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.590934] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.590934] env[61906]: DEBUG nova.network.neutron [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.710754] env[61906]: INFO nova.compute.manager [-] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Took 1.51 seconds to deallocate network for instance. [ 970.717049] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99251da9-da3b-4e35-a1db-b1de94f6344c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.724258] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8636df1-7156-41ec-9256-1dff440b75fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.759554] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc66407d-ccb0-4ae6-b909-d06e4c45a382 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.767929] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d11424-cdde-423d-a5e9-8b817bb64677 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.782497] env[61906]: DEBUG nova.compute.provider_tree [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.802961] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356984, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073174} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.803262] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 970.804067] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762bf7e7-b3b1-4a8d-8300-29603dadd7ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.826195] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] d2f16b79-c3fa-4450-8cb4-e9537215c60c/d2f16b79-c3fa-4450-8cb4-e9537215c60c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 970.826488] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d348821b-6128-47e9-8579-18eb3f1498c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.847010] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Waiting for the task: (returnval){ [ 970.847010] env[61906]: value = "task-1356985" [ 970.847010] env[61906]: _type = "Task" [ 970.847010] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.854485] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356985, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.907348] env[61906]: INFO nova.compute.manager [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Took 13.66 seconds to build instance. [ 971.220971] env[61906]: DEBUG oslo_concurrency.lockutils [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.286121] env[61906]: DEBUG nova.scheduler.client.report [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.300328] env[61906]: DEBUG nova.network.neutron [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance_info_cache with network_info: [{"id": "0c3f7a1f-6d39-4063-b093-29db43118645", "address": "fa:16:3e:77:db:95", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c3f7a1f-6d", "ovs_interfaceid": "0c3f7a1f-6d39-4063-b093-29db43118645", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.363954] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356985, 'name': ReconfigVM_Task, 'duration_secs': 0.515035} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.364286] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Reconfigured VM instance instance-0000005f to attach disk [datastore2] d2f16b79-c3fa-4450-8cb4-e9537215c60c/d2f16b79-c3fa-4450-8cb4-e9537215c60c.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 971.364940] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1bc67ece-1925-4f60-bd2b-c5ca5d3ab302 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.371457] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Waiting for the task: (returnval){ [ 971.371457] env[61906]: value = "task-1356986" [ 971.371457] env[61906]: _type = "Task" [ 971.371457] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.380078] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356986, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.409999] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d2e86050-acd1-4c48-893e-a85f3bc3e179 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e3578f5a-bd05-41fa-92e8-61d0e4ceed3f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.171s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.791623] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.713s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.795094] env[61906]: DEBUG oslo_concurrency.lockutils [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.574s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.795094] env[61906]: DEBUG nova.objects.instance [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lazy-loading 'resources' on Instance uuid 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.802941] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.813544] env[61906]: INFO nova.scheduler.client.report [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Deleted allocations for instance 99e5c55d-79ae-47a6-8500-79fc68291650 [ 971.881909] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356986, 'name': Rename_Task, 'duration_secs': 0.129272} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.882853] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 971.883180] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b0a7935-645a-4669-b1b1-516396217446 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.890288] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Waiting for the task: (returnval){ [ 971.890288] env[61906]: value = "task-1356987" [ 971.890288] env[61906]: _type = "Task" [ 971.890288] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.899060] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356987, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.128803] env[61906]: INFO nova.compute.manager [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Rebuilding instance [ 972.165928] env[61906]: DEBUG nova.compute.manager [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 972.166857] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a2edacf-03b6-42d7-b5be-c20c13ae14d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.324349] env[61906]: DEBUG oslo_concurrency.lockutils [None req-fb6133fb-6d94-4213-a88b-0ca189974ccf tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "99e5c55d-79ae-47a6-8500-79fc68291650" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.705s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.401906] env[61906]: DEBUG oslo_vmware.api [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356987, 'name': PowerOnVM_Task, 'duration_secs': 0.425626} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.402210] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 972.402423] env[61906]: INFO nova.compute.manager [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Took 7.15 seconds to spawn the instance on the hypervisor. [ 972.402607] env[61906]: DEBUG nova.compute.manager [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 972.403379] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e63309c-f1fd-477f-9ee7-5f689a762153 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.430182] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e25264-2533-4623-b15d-a6aa44f31535 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.438351] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-094f4600-3e5f-45b9-92c7-ff21e66f8fbc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.471086] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44fa0e91-20bc-4d7f-af02-0981b99cf5e9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.479983] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af018e31-a58d-4b1b-ba52-4f502b66c1d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.495753] env[61906]: DEBUG nova.compute.provider_tree [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.679421] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 972.679798] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b324a43-cb22-4ac8-a2e4-6dd8f4e780e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.687389] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 972.687389] env[61906]: value = "task-1356988" [ 972.687389] env[61906]: _type = "Task" [ 972.687389] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.695846] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356988, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.922620] env[61906]: INFO nova.compute.manager [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Took 13.42 seconds to build instance. [ 972.998650] env[61906]: DEBUG nova.scheduler.client.report [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 973.197882] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356988, 'name': PowerOffVM_Task, 'duration_secs': 0.347146} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.198250] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 973.198594] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 973.199254] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0790b4-7bf7-4246-bb58-2c35550de5cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.206035] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 973.206290] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e2459b2-ad8c-4065-b060-c66106adaf76 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.267729] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 973.267729] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 973.267975] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleting the datastore file [datastore2] e3578f5a-bd05-41fa-92e8-61d0e4ceed3f {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.268058] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-836a610a-4ae7-4a29-99ee-f74283a4d55f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.274491] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 973.274491] env[61906]: value = "task-1356990" [ 973.274491] env[61906]: _type = "Task" [ 973.274491] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.282341] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356990, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.320331] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ccdc719-4a9d-4238-a40f-a28e6994cc67 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.338084] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance '082b1a00-dd07-4d37-860c-6a91d6d40eaf' progress to 0 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 973.424933] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c20c4575-4a53-4422-9aa3-9c67a9118c70 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Lock "d2f16b79-c3fa-4450-8cb4-e9537215c60c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.927s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.503411] env[61906]: DEBUG oslo_concurrency.lockutils [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.709s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.529475] env[61906]: INFO nova.scheduler.client.report [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Deleted allocations for instance 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63 [ 973.788278] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1356990, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138076} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.790021] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.790021] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 973.790021] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 973.845129] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 973.845297] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3230232f-838e-4061-b871-76ce60ebe366 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.853671] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 973.853671] env[61906]: value = "task-1356991" [ 973.853671] env[61906]: _type = "Task" [ 973.853671] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.866429] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356991, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.041063] env[61906]: DEBUG oslo_concurrency.lockutils [None req-49fb8477-0c17-4473-80ae-614d0b62ed1c tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.047s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.227191] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Acquiring lock "d2f16b79-c3fa-4450-8cb4-e9537215c60c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.227586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Lock "d2f16b79-c3fa-4450-8cb4-e9537215c60c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.227767] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Acquiring lock "d2f16b79-c3fa-4450-8cb4-e9537215c60c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.227986] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Lock "d2f16b79-c3fa-4450-8cb4-e9537215c60c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.228208] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Lock "d2f16b79-c3fa-4450-8cb4-e9537215c60c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.230594] env[61906]: INFO nova.compute.manager [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Terminating instance [ 974.232569] env[61906]: DEBUG nova.compute.manager [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 974.232769] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 974.233627] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540a1037-7b56-41d8-84e4-adc1f5bcf263 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.241433] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 974.241678] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4bdf0d73-abc7-4de6-b8e5-9c9b2a95b4fd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.247951] env[61906]: DEBUG oslo_vmware.api [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Waiting for the task: (returnval){ [ 974.247951] env[61906]: value = "task-1356992" [ 974.247951] env[61906]: _type = "Task" [ 974.247951] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.255895] env[61906]: DEBUG oslo_vmware.api [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356992, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.364676] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356991, 'name': PowerOffVM_Task, 'duration_secs': 0.360837} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.364782] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 974.364974] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance '082b1a00-dd07-4d37-860c-6a91d6d40eaf' progress to 17 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 974.758332] env[61906]: DEBUG oslo_vmware.api [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356992, 'name': PowerOffVM_Task, 'duration_secs': 0.192712} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.758569] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 974.758753] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 974.759071] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0bc1b2d8-538c-4549-bcdd-85a8f33733c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.821428] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 974.821655] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 974.821840] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Deleting the datastore file [datastore2] d2f16b79-c3fa-4450-8cb4-e9537215c60c {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 974.822114] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fa217b99-2918-440a-b022-5ce3af8aa396 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.826708] env[61906]: DEBUG nova.virt.hardware [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 974.826941] env[61906]: DEBUG nova.virt.hardware [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 974.827117] env[61906]: DEBUG nova.virt.hardware [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.827312] env[61906]: DEBUG nova.virt.hardware [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 974.827462] env[61906]: DEBUG nova.virt.hardware [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.827614] env[61906]: DEBUG nova.virt.hardware [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 974.827834] env[61906]: DEBUG nova.virt.hardware [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 974.827996] env[61906]: DEBUG nova.virt.hardware [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 974.828190] env[61906]: DEBUG nova.virt.hardware [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 974.828363] env[61906]: DEBUG nova.virt.hardware [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 974.828590] env[61906]: DEBUG nova.virt.hardware [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.829662] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3d5a30-1d0e-4d35-8dc5-d50461a50e1b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.834031] env[61906]: DEBUG oslo_vmware.api [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Waiting for the task: (returnval){ [ 974.834031] env[61906]: value = "task-1356995" [ 974.834031] env[61906]: _type = "Task" [ 974.834031] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.840357] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b13edc-6220-451d-bca7-95076755df8e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.847029] env[61906]: DEBUG oslo_vmware.api [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356995, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.856836] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:6b:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8233b50c-be21-441a-a357-a29138a95b8b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f8b6d4d1-151c-45eb-b1c6-383bfaedef5b', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 974.864273] env[61906]: DEBUG oslo.service.loopingcall [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.864583] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 974.864829] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a6e84d2e-c494-4c45-8973-1b2879240a06 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.880348] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 974.880558] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 974.880678] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.880917] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 974.881079] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.881242] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 974.881456] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 974.881618] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 974.881788] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 974.881955] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 974.882161] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.886924] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c91dcac-a695-4ce0-be74-c00213fcdad0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.903474] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.903474] env[61906]: value = "task-1356996" [ 974.903474] env[61906]: _type = "Task" [ 974.903474] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.905207] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 974.905207] env[61906]: value = "task-1356997" [ 974.905207] env[61906]: _type = "Task" [ 974.905207] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.916526] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356996, 'name': CreateVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.922225] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356997, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.040013] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.040289] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.344975] env[61906]: DEBUG oslo_vmware.api [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Task: {'id': task-1356995, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230873} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.345359] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 975.345450] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 975.345628] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 975.345847] env[61906]: INFO nova.compute.manager [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 975.346256] env[61906]: DEBUG oslo.service.loopingcall [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 975.346489] env[61906]: DEBUG nova.compute.manager [-] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 975.346591] env[61906]: DEBUG nova.network.neutron [-] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 975.417090] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1356996, 'name': CreateVM_Task, 'duration_secs': 0.329838} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.419950] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 975.420271] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356997, 'name': ReconfigVM_Task, 'duration_secs': 0.21567} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.420869] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.421052] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.421375] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 975.421668] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance '082b1a00-dd07-4d37-860c-6a91d6d40eaf' progress to 33 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 975.424856] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1d9cb8e-1cb4-46ea-85ee-9ee10480d99f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.429272] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 975.429272] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52241e03-084f-02ad-24bc-9f2f6fab8e98" [ 975.429272] env[61906]: _type = "Task" [ 975.429272] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.436707] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52241e03-084f-02ad-24bc-9f2f6fab8e98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.543354] env[61906]: DEBUG nova.compute.manager [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 975.577875] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.578482] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.647815] env[61906]: DEBUG nova.compute.manager [req-aacf5a14-97f0-4570-8322-f5590763fdb8 req-d193c6e5-45e4-484c-ba80-ac6713e7e8ff service nova] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Received event network-vif-deleted-0ecf55ff-0f31-4c3b-9890-048efd2aa3a5 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.647815] env[61906]: INFO nova.compute.manager [req-aacf5a14-97f0-4570-8322-f5590763fdb8 req-d193c6e5-45e4-484c-ba80-ac6713e7e8ff service nova] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Neutron deleted interface 0ecf55ff-0f31-4c3b-9890-048efd2aa3a5; detaching it from the instance and deleting it from the info cache [ 975.647966] env[61906]: DEBUG nova.network.neutron [req-aacf5a14-97f0-4570-8322-f5590763fdb8 req-d193c6e5-45e4-484c-ba80-ac6713e7e8ff service nova] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.929242] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.929484] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.929542] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.929697] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.929862] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.930026] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.930236] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.930402] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.930576] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.930748] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.930962] env[61906]: DEBUG nova.virt.hardware [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.936469] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Reconfiguring VM instance instance-0000005d to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 975.936877] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63ec08bc-227c-4a7d-bca3-2c57ab5eba3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.959198] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52241e03-084f-02ad-24bc-9f2f6fab8e98, 'name': SearchDatastore_Task, 'duration_secs': 0.018814} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.959858] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.960165] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 975.960369] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.960515] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.960703] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 975.961607] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-011e4318-7f17-4aa6-8a9d-0b654bf3ca35 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.963264] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 975.963264] env[61906]: value = "task-1356998" [ 975.963264] env[61906]: _type = "Task" [ 975.963264] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.970998] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356998, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.972069] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 975.972254] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 975.972898] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b73d737-be3c-4205-a181-0becae19e489 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.977243] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 975.977243] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522be39c-2470-7ff4-7de9-79aa3b3fab43" [ 975.977243] env[61906]: _type = "Task" [ 975.977243] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.984683] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522be39c-2470-7ff4-7de9-79aa3b3fab43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.066746] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.067040] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.068552] env[61906]: INFO nova.compute.claims [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 976.081036] env[61906]: INFO nova.compute.manager [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Detaching volume 4d5078bb-42a4-4eaa-b253-4e94f1321e51 [ 976.113329] env[61906]: INFO nova.virt.block_device [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Attempting to driver detach volume 4d5078bb-42a4-4eaa-b253-4e94f1321e51 from mountpoint /dev/sdb [ 976.113565] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Volume detach. Driver type: vmdk {{(pid=61906) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 976.113756] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-289016', 'volume_id': '4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'name': 'volume-4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'cb879dc0-3af7-4279-aa28-66e2b8b4286d', 'attached_at': '', 'detached_at': '', 'volume_id': '4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'serial': '4d5078bb-42a4-4eaa-b253-4e94f1321e51'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 976.114659] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0225754e-76f2-47b1-99a0-e92a8b221466 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.137649] env[61906]: DEBUG nova.network.neutron [-] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.139654] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27df4694-7fb1-42e5-9aa0-534b24173bce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.147729] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4defa15-c866-4e6d-a443-5694412bcdb8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.151173] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-75fd5020-d0f3-4e09-8e5b-23512f2e6a83 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.173798] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a481f452-3723-48ac-be9e-3c86576e6288 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.179413] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38d9599-0c81-4242-a8f3-641abbe088c7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.205083] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] The volume has not been displaced from its original location: [datastore2] volume-4d5078bb-42a4-4eaa-b253-4e94f1321e51/volume-4d5078bb-42a4-4eaa-b253-4e94f1321e51.vmdk. No consolidation needed. {{(pid=61906) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 976.210828] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Reconfiguring VM instance instance-00000049 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 976.218558] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b915c033-cc3d-4ad9-8c6b-94a6861f42b3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.232759] env[61906]: DEBUG nova.compute.manager [req-aacf5a14-97f0-4570-8322-f5590763fdb8 req-d193c6e5-45e4-484c-ba80-ac6713e7e8ff service nova] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Detach interface failed, port_id=0ecf55ff-0f31-4c3b-9890-048efd2aa3a5, reason: Instance d2f16b79-c3fa-4450-8cb4-e9537215c60c could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 976.238107] env[61906]: DEBUG oslo_vmware.api [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 976.238107] env[61906]: value = "task-1356999" [ 976.238107] env[61906]: _type = "Task" [ 976.238107] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.246080] env[61906]: DEBUG oslo_vmware.api [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356999, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.472825] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1356998, 'name': ReconfigVM_Task, 'duration_secs': 0.148411} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.473145] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Reconfigured VM instance instance-0000005d to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 976.474038] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ed9103-2ee3-4b10-a66a-4e1a4b956b73 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.497357] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 082b1a00-dd07-4d37-860c-6a91d6d40eaf/082b1a00-dd07-4d37-860c-6a91d6d40eaf.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.497915] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ec7579a-b4e7-4419-aefa-1858fc68c221 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.513239] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522be39c-2470-7ff4-7de9-79aa3b3fab43, 'name': SearchDatastore_Task, 'duration_secs': 0.017017} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.514261] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29264461-ced7-46ff-b8ce-c6499b87f313 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.517418] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 976.517418] env[61906]: value = "task-1357000" [ 976.517418] env[61906]: _type = "Task" [ 976.517418] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.520974] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 976.520974] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c24f70-9abd-8fa3-cb63-97e096e8a090" [ 976.520974] env[61906]: _type = "Task" [ 976.520974] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.526901] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1357000, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.531480] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c24f70-9abd-8fa3-cb63-97e096e8a090, 'name': SearchDatastore_Task, 'duration_secs': 0.009598} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.531707] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.531959] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] e3578f5a-bd05-41fa-92e8-61d0e4ceed3f/e3578f5a-bd05-41fa-92e8-61d0e4ceed3f.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 976.532205] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5557a57f-8f9c-4fd6-82d1-3c88f0084979 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.538043] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 976.538043] env[61906]: value = "task-1357001" [ 976.538043] env[61906]: _type = "Task" [ 976.538043] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.544960] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357001, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.644274] env[61906]: INFO nova.compute.manager [-] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Took 1.30 seconds to deallocate network for instance. [ 976.749467] env[61906]: DEBUG oslo_vmware.api [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1356999, 'name': ReconfigVM_Task, 'duration_secs': 0.215592} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.749775] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Reconfigured VM instance instance-00000049 to detach disk 2001 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 976.754569] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae940897-32f4-4c24-995d-207ccb5c898f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.770224] env[61906]: DEBUG oslo_vmware.api [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 976.770224] env[61906]: value = "task-1357002" [ 976.770224] env[61906]: _type = "Task" [ 976.770224] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.779374] env[61906]: DEBUG oslo_vmware.api [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357002, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.028858] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1357000, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.046972] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357001, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481448} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.047253] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] e3578f5a-bd05-41fa-92e8-61d0e4ceed3f/e3578f5a-bd05-41fa-92e8-61d0e4ceed3f.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 977.047468] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 977.047737] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f1464e7-f62d-4d3c-9086-4eca8fdad7a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.053243] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 977.053243] env[61906]: value = "task-1357003" [ 977.053243] env[61906]: _type = "Task" [ 977.053243] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.059993] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357003, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.150883] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.183417] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf87815-eecd-4443-9579-3f0cd898e2e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.190748] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6effe871-28bc-44e9-9015-76b81f879d7f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.220051] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7507716e-ca09-42c6-89f5-8206a54acbfa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.226748] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5660810-3ca9-4eb5-b27d-bf0552c193c5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.239387] env[61906]: DEBUG nova.compute.provider_tree [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.279677] env[61906]: DEBUG oslo_vmware.api [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357002, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.527826] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1357000, 'name': ReconfigVM_Task, 'duration_secs': 0.555792} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.528174] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 082b1a00-dd07-4d37-860c-6a91d6d40eaf/082b1a00-dd07-4d37-860c-6a91d6d40eaf.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.528424] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance '082b1a00-dd07-4d37-860c-6a91d6d40eaf' progress to 50 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 977.562265] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357003, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057583} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.562490] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 977.563265] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7022c808-a34d-44d7-bb90-6142ed6b9384 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.585962] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] e3578f5a-bd05-41fa-92e8-61d0e4ceed3f/e3578f5a-bd05-41fa-92e8-61d0e4ceed3f.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.586476] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8c0b3e9-b868-4a35-8dd7-169a8d03b355 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.605192] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 977.605192] env[61906]: value = "task-1357004" [ 977.605192] env[61906]: _type = "Task" [ 977.605192] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.612644] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357004, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.742928] env[61906]: DEBUG nova.scheduler.client.report [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 977.782336] env[61906]: DEBUG oslo_vmware.api [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357002, 'name': ReconfigVM_Task, 'duration_secs': 0.746673} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.782667] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-289016', 'volume_id': '4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'name': 'volume-4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'cb879dc0-3af7-4279-aa28-66e2b8b4286d', 'attached_at': '', 'detached_at': '', 'volume_id': '4d5078bb-42a4-4eaa-b253-4e94f1321e51', 'serial': '4d5078bb-42a4-4eaa-b253-4e94f1321e51'} {{(pid=61906) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 978.034634] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077e458b-bf00-43d5-976c-bf204ab7f143 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.053761] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6701065b-1522-4556-a886-cd9eb61ffdcb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.070555] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance '082b1a00-dd07-4d37-860c-6a91d6d40eaf' progress to 67 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 978.117033] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357004, 'name': ReconfigVM_Task, 'duration_secs': 0.246633} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.117033] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Reconfigured VM instance instance-0000005e to attach disk [datastore2] e3578f5a-bd05-41fa-92e8-61d0e4ceed3f/e3578f5a-bd05-41fa-92e8-61d0e4ceed3f.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.117619] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1813ef84-7491-488e-aa7d-1871621ac5f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.123901] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 978.123901] env[61906]: value = "task-1357005" [ 978.123901] env[61906]: _type = "Task" [ 978.123901] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.131495] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357005, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.247986] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.181s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.248945] env[61906]: DEBUG nova.compute.manager [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 978.251873] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.101s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.252108] env[61906]: DEBUG nova.objects.instance [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Lazy-loading 'resources' on Instance uuid d2f16b79-c3fa-4450-8cb4-e9537215c60c {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.323197] env[61906]: DEBUG nova.objects.instance [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lazy-loading 'flavor' on Instance uuid cb879dc0-3af7-4279-aa28-66e2b8b4286d {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.629284] env[61906]: DEBUG nova.network.neutron [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Port 0c3f7a1f-6d39-4063-b093-29db43118645 binding to destination host cpu-1 is already ACTIVE {{(pid=61906) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 978.636243] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357005, 'name': Rename_Task, 'duration_secs': 0.12836} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.636519] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 978.636861] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da1b47b2-6ee9-423b-bfe0-ff7bd3a5512f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.643644] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 978.643644] env[61906]: value = "task-1357006" [ 978.643644] env[61906]: _type = "Task" [ 978.643644] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.651413] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357006, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.755403] env[61906]: DEBUG nova.compute.utils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 978.759560] env[61906]: DEBUG nova.compute.manager [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 978.759778] env[61906]: DEBUG nova.network.neutron [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 978.798140] env[61906]: DEBUG nova.policy [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73219f5e96f64d1da86a19da9b18b06a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5533be74874b4094b7ef0f98a121f60b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 978.883899] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb7ccad-9f28-48b3-b0fd-8e272c466adf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.891732] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01a7994-30f5-4c09-8aed-bd7acbd2efae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.922046] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7cebc3-7337-41cf-93eb-91e1509874d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.929216] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519cccc6-3d87-495d-b822-702ed52faedf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.944560] env[61906]: DEBUG nova.compute.provider_tree [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.081080] env[61906]: DEBUG nova.network.neutron [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Successfully created port: fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 979.153503] env[61906]: DEBUG oslo_vmware.api [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357006, 'name': PowerOnVM_Task, 'duration_secs': 0.425257} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.153503] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 979.153670] env[61906]: DEBUG nova.compute.manager [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 979.154315] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad47577-9cce-4c10-84a7-af3cf2f1469a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.260228] env[61906]: DEBUG nova.compute.manager [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 979.329848] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d89fa0f0-b47d-4a52-bba3-46af49c37ed6 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.751s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.363765] env[61906]: DEBUG oslo_concurrency.lockutils [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "7bae658e-b050-4639-b34e-c2671ef5c773" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.364046] env[61906]: DEBUG oslo_concurrency.lockutils [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7bae658e-b050-4639-b34e-c2671ef5c773" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.364260] env[61906]: DEBUG oslo_concurrency.lockutils [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "7bae658e-b050-4639-b34e-c2671ef5c773-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.364448] env[61906]: DEBUG oslo_concurrency.lockutils [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7bae658e-b050-4639-b34e-c2671ef5c773-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.364624] env[61906]: DEBUG oslo_concurrency.lockutils [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7bae658e-b050-4639-b34e-c2671ef5c773-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.366700] env[61906]: INFO nova.compute.manager [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Terminating instance [ 979.368432] env[61906]: DEBUG nova.compute.manager [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 979.368629] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 979.369516] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524b90b7-8cf5-4a2e-81c8-2804937e4c98 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.376416] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 979.376646] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ea11918-1774-4d9a-bb28-1e081610772b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.383149] env[61906]: DEBUG oslo_vmware.api [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 979.383149] env[61906]: value = "task-1357007" [ 979.383149] env[61906]: _type = "Task" [ 979.383149] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.390231] env[61906]: DEBUG oslo_vmware.api [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1357007, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.447478] env[61906]: DEBUG nova.scheduler.client.report [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.647684] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.648026] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.648129] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.669122] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.838497] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.838803] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.839047] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.839256] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.839433] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.841563] env[61906]: INFO nova.compute.manager [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Terminating instance [ 979.843310] env[61906]: DEBUG nova.compute.manager [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 979.843507] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 979.844396] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f98477-6ee7-4d7e-98bc-f4ca317388b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.853784] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 979.854038] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14d54ded-affd-4bb6-8225-18cc93a4494c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.860141] env[61906]: DEBUG oslo_vmware.api [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 979.860141] env[61906]: value = "task-1357008" [ 979.860141] env[61906]: _type = "Task" [ 979.860141] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.868275] env[61906]: DEBUG oslo_vmware.api [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357008, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.892886] env[61906]: DEBUG oslo_vmware.api [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1357007, 'name': PowerOffVM_Task, 'duration_secs': 0.194853} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.893190] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 979.893371] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 979.893713] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7bad07a0-3c8d-4b8e-b151-c769a415029c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.952830] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.701s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.955338] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.286s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.955505] env[61906]: DEBUG nova.objects.instance [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61906) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 979.970517] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 979.970746] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 979.970936] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Deleting the datastore file [datastore1] 7bae658e-b050-4639-b34e-c2671ef5c773 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.971221] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-484812eb-e1dd-4b95-af78-0d382a2f9d3b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.976570] env[61906]: INFO nova.scheduler.client.report [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Deleted allocations for instance d2f16b79-c3fa-4450-8cb4-e9537215c60c [ 979.979319] env[61906]: DEBUG oslo_vmware.api [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 979.979319] env[61906]: value = "task-1357010" [ 979.979319] env[61906]: _type = "Task" [ 979.979319] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.990035] env[61906]: DEBUG oslo_vmware.api [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1357010, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.270232] env[61906]: DEBUG nova.compute.manager [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 980.303053] env[61906]: DEBUG nova.virt.hardware [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 980.303386] env[61906]: DEBUG nova.virt.hardware [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 980.303575] env[61906]: DEBUG nova.virt.hardware [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 980.303770] env[61906]: DEBUG nova.virt.hardware [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 980.303973] env[61906]: DEBUG nova.virt.hardware [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 980.304185] env[61906]: DEBUG nova.virt.hardware [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 980.304410] env[61906]: DEBUG nova.virt.hardware [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 980.304576] env[61906]: DEBUG nova.virt.hardware [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 980.304777] env[61906]: DEBUG nova.virt.hardware [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 980.304956] env[61906]: DEBUG nova.virt.hardware [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 980.305150] env[61906]: DEBUG nova.virt.hardware [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 980.306061] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd4f282-3003-4221-9f13-44a10c5633a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.314819] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-580d733e-0f1d-4fdd-9b70-8c0b20396dcb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.368957] env[61906]: DEBUG oslo_vmware.api [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357008, 'name': PowerOffVM_Task, 'duration_secs': 0.185587} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.369258] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 980.369427] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 980.369682] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9be311a9-b4c6-42c5-81f4-c71c181b5d63 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.405163] env[61906]: DEBUG oslo_concurrency.lockutils [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "e3578f5a-bd05-41fa-92e8-61d0e4ceed3f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.405831] env[61906]: DEBUG oslo_concurrency.lockutils [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e3578f5a-bd05-41fa-92e8-61d0e4ceed3f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.406070] env[61906]: DEBUG oslo_concurrency.lockutils [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "e3578f5a-bd05-41fa-92e8-61d0e4ceed3f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.406273] env[61906]: DEBUG oslo_concurrency.lockutils [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e3578f5a-bd05-41fa-92e8-61d0e4ceed3f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.406449] env[61906]: DEBUG oslo_concurrency.lockutils [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e3578f5a-bd05-41fa-92e8-61d0e4ceed3f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.408652] env[61906]: INFO nova.compute.manager [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Terminating instance [ 980.410506] env[61906]: DEBUG nova.compute.manager [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 980.410701] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 980.411561] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f01db1-199e-4449-8091-008314bdc5e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.418775] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 980.419373] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9cdf58dd-3005-4807-9998-2a7563322f9c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.425534] env[61906]: DEBUG oslo_vmware.api [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 980.425534] env[61906]: value = "task-1357012" [ 980.425534] env[61906]: _type = "Task" [ 980.425534] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.434288] env[61906]: DEBUG oslo_vmware.api [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357012, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.444775] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 980.445048] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 980.445247] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleting the datastore file [datastore1] cb879dc0-3af7-4279-aa28-66e2b8b4286d {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 980.445542] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d12ad14-6208-46cc-b76b-28a80b869ed9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.451669] env[61906]: DEBUG oslo_vmware.api [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 980.451669] env[61906]: value = "task-1357013" [ 980.451669] env[61906]: _type = "Task" [ 980.451669] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.463407] env[61906]: DEBUG oslo_vmware.api [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357013, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.491583] env[61906]: DEBUG oslo_concurrency.lockutils [None req-41009ba1-93e5-4fb9-a561-15b6e28d9820 tempest-ServerGroupTestJSON-2059420664 tempest-ServerGroupTestJSON-2059420664-project-member] Lock "d2f16b79-c3fa-4450-8cb4-e9537215c60c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.264s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.496021] env[61906]: DEBUG oslo_vmware.api [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1357010, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.221449} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.496283] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.496544] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 980.496751] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 980.496930] env[61906]: INFO nova.compute.manager [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Took 1.13 seconds to destroy the instance on the hypervisor. [ 980.497192] env[61906]: DEBUG oslo.service.loopingcall [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.497400] env[61906]: DEBUG nova.compute.manager [-] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 980.497492] env[61906]: DEBUG nova.network.neutron [-] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 980.525345] env[61906]: DEBUG nova.compute.manager [req-5aef4973-e4e0-446f-b35c-2f8269ad3ed2 req-7fa33d23-08a1-408e-89df-f1d0cb3feeb7 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Received event network-vif-plugged-fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.525511] env[61906]: DEBUG oslo_concurrency.lockutils [req-5aef4973-e4e0-446f-b35c-2f8269ad3ed2 req-7fa33d23-08a1-408e-89df-f1d0cb3feeb7 service nova] Acquiring lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.525761] env[61906]: DEBUG oslo_concurrency.lockutils [req-5aef4973-e4e0-446f-b35c-2f8269ad3ed2 req-7fa33d23-08a1-408e-89df-f1d0cb3feeb7 service nova] Lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.525939] env[61906]: DEBUG oslo_concurrency.lockutils [req-5aef4973-e4e0-446f-b35c-2f8269ad3ed2 req-7fa33d23-08a1-408e-89df-f1d0cb3feeb7 service nova] Lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.526122] env[61906]: DEBUG nova.compute.manager [req-5aef4973-e4e0-446f-b35c-2f8269ad3ed2 req-7fa33d23-08a1-408e-89df-f1d0cb3feeb7 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] No waiting events found dispatching network-vif-plugged-fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 980.526292] env[61906]: WARNING nova.compute.manager [req-5aef4973-e4e0-446f-b35c-2f8269ad3ed2 req-7fa33d23-08a1-408e-89df-f1d0cb3feeb7 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Received unexpected event network-vif-plugged-fc9835ce-e30c-4b58-8804-1062f5d166de for instance with vm_state building and task_state spawning. [ 980.721397] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.721721] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.721792] env[61906]: DEBUG nova.network.neutron [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 980.797298] env[61906]: DEBUG nova.network.neutron [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Successfully updated port: fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 980.937908] env[61906]: DEBUG oslo_vmware.api [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357012, 'name': PowerOffVM_Task, 'duration_secs': 0.222807} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.938159] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 980.938468] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 980.938663] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e2ae4a8-d788-4611-8582-eb8f9ff87bec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.961948] env[61906]: DEBUG oslo_vmware.api [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357013, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.276346} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.962249] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.962439] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 980.962611] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 980.962808] env[61906]: INFO nova.compute.manager [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 980.963710] env[61906]: DEBUG oslo.service.loopingcall [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.966564] env[61906]: DEBUG nova.compute.manager [-] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 980.966726] env[61906]: DEBUG nova.network.neutron [-] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 980.975109] env[61906]: DEBUG oslo_concurrency.lockutils [None req-ed8aebf9-275a-48c7-8f6e-4fd33916df4f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.011060] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.011304] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.011492] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleting the datastore file [datastore2] e3578f5a-bd05-41fa-92e8-61d0e4ceed3f {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.011765] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b30ea5d8-e9fb-45b8-950e-014f5a73b698 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.018758] env[61906]: DEBUG oslo_vmware.api [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 981.018758] env[61906]: value = "task-1357015" [ 981.018758] env[61906]: _type = "Task" [ 981.018758] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.029477] env[61906]: DEBUG oslo_vmware.api [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357015, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.299968] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.301024] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.301234] env[61906]: DEBUG nova.network.neutron [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.477341] env[61906]: DEBUG nova.network.neutron [-] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.530182] env[61906]: DEBUG oslo_vmware.api [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357015, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.215194} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.531500] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.532081] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 981.532445] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 981.532650] env[61906]: INFO nova.compute.manager [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 981.533098] env[61906]: DEBUG oslo.service.loopingcall [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.534402] env[61906]: DEBUG nova.compute.manager [-] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 981.534402] env[61906]: DEBUG nova.network.neutron [-] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 981.579329] env[61906]: DEBUG nova.network.neutron [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance_info_cache with network_info: [{"id": "0c3f7a1f-6d39-4063-b093-29db43118645", "address": "fa:16:3e:77:db:95", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c3f7a1f-6d", "ovs_interfaceid": "0c3f7a1f-6d39-4063-b093-29db43118645", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.872225] env[61906]: DEBUG nova.network.neutron [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 981.937034] env[61906]: DEBUG nova.compute.manager [req-3b348db7-5e4c-4f21-b917-98a899ee852c req-1bb62146-d3c7-470f-8623-5eb6d6147230 service nova] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Received event network-vif-deleted-f8b6d4d1-151c-45eb-b1c6-383bfaedef5b {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.937212] env[61906]: INFO nova.compute.manager [req-3b348db7-5e4c-4f21-b917-98a899ee852c req-1bb62146-d3c7-470f-8623-5eb6d6147230 service nova] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Neutron deleted interface f8b6d4d1-151c-45eb-b1c6-383bfaedef5b; detaching it from the instance and deleting it from the info cache [ 981.937393] env[61906]: DEBUG nova.network.neutron [req-3b348db7-5e4c-4f21-b917-98a899ee852c req-1bb62146-d3c7-470f-8623-5eb6d6147230 service nova] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.982313] env[61906]: INFO nova.compute.manager [-] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Took 1.48 seconds to deallocate network for instance. [ 982.063558] env[61906]: DEBUG nova.network.neutron [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updating instance_info_cache with network_info: [{"id": "fc9835ce-e30c-4b58-8804-1062f5d166de", "address": "fa:16:3e:b8:59:0b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc9835ce-e3", "ovs_interfaceid": "fc9835ce-e30c-4b58-8804-1062f5d166de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.085809] env[61906]: DEBUG oslo_concurrency.lockutils [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.119761] env[61906]: DEBUG nova.network.neutron [-] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.402818] env[61906]: DEBUG nova.network.neutron [-] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.441331] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7278b404-303a-46a9-8e21-5b9c5c2d01d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.450717] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751bff97-e499-47c9-96d2-5b1181e5b465 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.475402] env[61906]: DEBUG nova.compute.manager [req-3b348db7-5e4c-4f21-b917-98a899ee852c req-1bb62146-d3c7-470f-8623-5eb6d6147230 service nova] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Detach interface failed, port_id=f8b6d4d1-151c-45eb-b1c6-383bfaedef5b, reason: Instance e3578f5a-bd05-41fa-92e8-61d0e4ceed3f could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 982.489609] env[61906]: DEBUG oslo_concurrency.lockutils [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.489898] env[61906]: DEBUG oslo_concurrency.lockutils [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.490204] env[61906]: DEBUG nova.objects.instance [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lazy-loading 'resources' on Instance uuid 7bae658e-b050-4639-b34e-c2671ef5c773 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.556886] env[61906]: DEBUG nova.compute.manager [req-7e76a13a-565a-4470-89d3-2dabcb4593fb req-1cb29cc2-92f5-4973-8088-6c60b98e4398 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Received event network-changed-fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.557111] env[61906]: DEBUG nova.compute.manager [req-7e76a13a-565a-4470-89d3-2dabcb4593fb req-1cb29cc2-92f5-4973-8088-6c60b98e4398 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Refreshing instance network info cache due to event network-changed-fc9835ce-e30c-4b58-8804-1062f5d166de. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.557483] env[61906]: DEBUG oslo_concurrency.lockutils [req-7e76a13a-565a-4470-89d3-2dabcb4593fb req-1cb29cc2-92f5-4973-8088-6c60b98e4398 service nova] Acquiring lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.568039] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.568546] env[61906]: DEBUG nova.compute.manager [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Instance network_info: |[{"id": "fc9835ce-e30c-4b58-8804-1062f5d166de", "address": "fa:16:3e:b8:59:0b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc9835ce-e3", "ovs_interfaceid": "fc9835ce-e30c-4b58-8804-1062f5d166de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 982.568773] env[61906]: DEBUG oslo_concurrency.lockutils [req-7e76a13a-565a-4470-89d3-2dabcb4593fb req-1cb29cc2-92f5-4973-8088-6c60b98e4398 service nova] Acquired lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.568852] env[61906]: DEBUG nova.network.neutron [req-7e76a13a-565a-4470-89d3-2dabcb4593fb req-1cb29cc2-92f5-4973-8088-6c60b98e4398 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Refreshing network info cache for port fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 982.570364] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:59:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fc9835ce-e30c-4b58-8804-1062f5d166de', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.577589] env[61906]: DEBUG oslo.service.loopingcall [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.578139] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 982.578860] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b128397c-5de6-497a-817f-af9dca5c45b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.602660] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.602660] env[61906]: value = "task-1357016" [ 982.602660] env[61906]: _type = "Task" [ 982.602660] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.610735] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357016, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.613748] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3174eb7-1469-4411-bbba-17d496224a1b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.630843] env[61906]: INFO nova.compute.manager [-] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Took 1.66 seconds to deallocate network for instance. [ 982.633575] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ef41b5-eac7-4884-9cdb-9ce466af1e98 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.642763] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance '082b1a00-dd07-4d37-860c-6a91d6d40eaf' progress to 83 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 982.905547] env[61906]: INFO nova.compute.manager [-] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Took 1.37 seconds to deallocate network for instance. [ 983.082667] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-facb4b19-3641-40a1-b7d0-bda34b7791d9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.090775] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b84912-c88e-4974-b4da-fd7044962349 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.125009] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f1e0c6-284d-4225-b880-6b34bc2012af {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.132403] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357016, 'name': CreateVM_Task, 'duration_secs': 0.326764} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.134278] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 983.134988] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.135179] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.135499] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 983.136670] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45428ed-b4d9-498e-a914-3583120b87f8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.140358] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5de03e25-580d-4897-b93b-13438b3cc7d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.151964] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.153815] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 983.153985] env[61906]: DEBUG nova.compute.provider_tree [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.156586] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19c5b95f-b3f3-45db-94e1-c8dd04d12243 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.158022] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 983.158022] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f56401-08c1-3d0f-bde3-aff1f64e6469" [ 983.158022] env[61906]: _type = "Task" [ 983.158022] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.162848] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 983.162848] env[61906]: value = "task-1357017" [ 983.162848] env[61906]: _type = "Task" [ 983.162848] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.167028] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f56401-08c1-3d0f-bde3-aff1f64e6469, 'name': SearchDatastore_Task, 'duration_secs': 0.011744} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.169802] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.170071] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.170314] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.170467] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.170650] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.171168] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2334bc0-d157-4563-8263-35cd089f5fbd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.181137] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1357017, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.182597] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.182782] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 983.183486] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bde1a5d7-db12-4c25-be1b-4580968fe2b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.188072] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 983.188072] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]524face1-1282-4cc5-f5c1-8664e83f5d8a" [ 983.188072] env[61906]: _type = "Task" [ 983.188072] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.195751] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]524face1-1282-4cc5-f5c1-8664e83f5d8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.346655] env[61906]: DEBUG nova.network.neutron [req-7e76a13a-565a-4470-89d3-2dabcb4593fb req-1cb29cc2-92f5-4973-8088-6c60b98e4398 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updated VIF entry in instance network info cache for port fc9835ce-e30c-4b58-8804-1062f5d166de. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 983.347083] env[61906]: DEBUG nova.network.neutron [req-7e76a13a-565a-4470-89d3-2dabcb4593fb req-1cb29cc2-92f5-4973-8088-6c60b98e4398 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updating instance_info_cache with network_info: [{"id": "fc9835ce-e30c-4b58-8804-1062f5d166de", "address": "fa:16:3e:b8:59:0b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc9835ce-e3", "ovs_interfaceid": "fc9835ce-e30c-4b58-8804-1062f5d166de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.412238] env[61906]: DEBUG oslo_concurrency.lockutils [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.661123] env[61906]: DEBUG nova.scheduler.client.report [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.675701] env[61906]: DEBUG oslo_vmware.api [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1357017, 'name': PowerOnVM_Task, 'duration_secs': 0.417845} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.676562] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 983.676852] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-dcda30c1-f4ab-451d-89b5-d941c824c06b tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance '082b1a00-dd07-4d37-860c-6a91d6d40eaf' progress to 100 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 983.698013] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]524face1-1282-4cc5-f5c1-8664e83f5d8a, 'name': SearchDatastore_Task, 'duration_secs': 0.009114} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.698812] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de8cc987-12a1-4b88-afb1-00dc4b841f12 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.703628] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 983.703628] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5283b2b1-b9ad-a2d1-983d-54ee1fb9b1a9" [ 983.703628] env[61906]: _type = "Task" [ 983.703628] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.710767] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5283b2b1-b9ad-a2d1-983d-54ee1fb9b1a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.850038] env[61906]: DEBUG oslo_concurrency.lockutils [req-7e76a13a-565a-4470-89d3-2dabcb4593fb req-1cb29cc2-92f5-4973-8088-6c60b98e4398 service nova] Releasing lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.850325] env[61906]: DEBUG nova.compute.manager [req-7e76a13a-565a-4470-89d3-2dabcb4593fb req-1cb29cc2-92f5-4973-8088-6c60b98e4398 service nova] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Received event network-vif-deleted-4c806e0c-9874-4943-9a71-05fc2c40045e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.850520] env[61906]: DEBUG nova.compute.manager [req-7e76a13a-565a-4470-89d3-2dabcb4593fb req-1cb29cc2-92f5-4973-8088-6c60b98e4398 service nova] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Received event network-vif-deleted-f119e802-0a46-4390-927d-805c2e23a4b4 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.165169] env[61906]: DEBUG oslo_concurrency.lockutils [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.675s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.167464] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.016s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.167699] env[61906]: DEBUG nova.objects.instance [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lazy-loading 'resources' on Instance uuid cb879dc0-3af7-4279-aa28-66e2b8b4286d {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.186424] env[61906]: INFO nova.scheduler.client.report [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Deleted allocations for instance 7bae658e-b050-4639-b34e-c2671ef5c773 [ 984.214104] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5283b2b1-b9ad-a2d1-983d-54ee1fb9b1a9, 'name': SearchDatastore_Task, 'duration_secs': 0.010437} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.214337] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.214591] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 3ea24d58-ffbd-41d4-a010-978a395ad2fc/3ea24d58-ffbd-41d4-a010-978a395ad2fc.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 984.214839] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a3fd748f-d1e9-4365-ab74-6a4ca5a215c1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.221802] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 984.221802] env[61906]: value = "task-1357018" [ 984.221802] env[61906]: _type = "Task" [ 984.221802] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.229010] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357018, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.694347] env[61906]: DEBUG oslo_concurrency.lockutils [None req-af4788b8-dc38-4e51-9111-48d67984bcff tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "7bae658e-b050-4639-b34e-c2671ef5c773" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.330s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.736461] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357018, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.774131] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689f0f12-6065-48d8-ba97-d3b4ae77f77f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.781428] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77aee49-94f9-4a32-a068-aab0a5afd0fd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.810582] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b81b67f-084e-487d-97fe-bbdd69e7aca1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.817388] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-687f1647-a0a7-4526-b6ac-f861940e2e27 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.830864] env[61906]: DEBUG nova.compute.provider_tree [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.928546] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "79452791-59cb-4722-bb4a-8e59d8c4e641" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.928844] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "79452791-59cb-4722-bb4a-8e59d8c4e641" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.929095] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "79452791-59cb-4722-bb4a-8e59d8c4e641-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.929294] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "79452791-59cb-4722-bb4a-8e59d8c4e641-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.929468] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "79452791-59cb-4722-bb4a-8e59d8c4e641-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.931521] env[61906]: INFO nova.compute.manager [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Terminating instance [ 984.933404] env[61906]: DEBUG nova.compute.manager [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 984.933612] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 984.934492] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfee2df-47ab-42a8-abad-0fa37f4808a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.942579] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.942811] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ba83520-7e57-4048-b52e-cfec59c94b9c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.948674] env[61906]: DEBUG oslo_vmware.api [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 984.948674] env[61906]: value = "task-1357019" [ 984.948674] env[61906]: _type = "Task" [ 984.948674] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.955868] env[61906]: DEBUG oslo_vmware.api [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1357019, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.234389] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357018, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51994} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.234758] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 3ea24d58-ffbd-41d4-a010-978a395ad2fc/3ea24d58-ffbd-41d4-a010-978a395ad2fc.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 985.234865] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 985.235067] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-88e17084-2555-4cdb-992e-0385708957cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.241369] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 985.241369] env[61906]: value = "task-1357020" [ 985.241369] env[61906]: _type = "Task" [ 985.241369] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.248645] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357020, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.334512] env[61906]: DEBUG nova.scheduler.client.report [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.458573] env[61906]: DEBUG oslo_vmware.api [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1357019, 'name': PowerOffVM_Task, 'duration_secs': 0.213629} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.458897] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 985.459111] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 985.459372] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b1e35224-f7e6-49f3-aa7b-325ea734699f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.525146] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 985.525363] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 985.525651] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Deleting the datastore file [datastore2] 79452791-59cb-4722-bb4a-8e59d8c4e641 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 985.526010] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dce8d9b5-2c37-49af-aa92-a15a3d50527f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.536617] env[61906]: DEBUG oslo_vmware.api [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for the task: (returnval){ [ 985.536617] env[61906]: value = "task-1357022" [ 985.536617] env[61906]: _type = "Task" [ 985.536617] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.545225] env[61906]: DEBUG oslo_vmware.api [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1357022, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.750850] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357020, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.145175} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.752031] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 985.752250] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ff9090-cf75-4a5a-a42f-5cc94525e39e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.774040] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 3ea24d58-ffbd-41d4-a010-978a395ad2fc/3ea24d58-ffbd-41d4-a010-978a395ad2fc.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.774351] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-352624df-8d1f-4014-aa6b-c9acc5d87c4c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.794530] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 985.794530] env[61906]: value = "task-1357023" [ 985.794530] env[61906]: _type = "Task" [ 985.794530] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.802685] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357023, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.840245] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.673s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.843411] env[61906]: DEBUG oslo_concurrency.lockutils [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.430s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.843695] env[61906]: DEBUG nova.objects.instance [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lazy-loading 'resources' on Instance uuid e3578f5a-bd05-41fa-92e8-61d0e4ceed3f {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.868231] env[61906]: INFO nova.scheduler.client.report [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleted allocations for instance cb879dc0-3af7-4279-aa28-66e2b8b4286d [ 986.038391] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.038645] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.038940] env[61906]: DEBUG nova.compute.manager [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Going to confirm migration 2 {{(pid=61906) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 986.049279] env[61906]: DEBUG oslo_vmware.api [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Task: {'id': task-1357022, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205332} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.049539] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.049724] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 986.049906] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 986.050099] env[61906]: INFO nova.compute.manager [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Took 1.12 seconds to destroy the instance on the hypervisor. [ 986.050351] env[61906]: DEBUG oslo.service.loopingcall [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.050917] env[61906]: DEBUG nova.compute.manager [-] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 986.050917] env[61906]: DEBUG nova.network.neutron [-] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 986.320077] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357023, 'name': ReconfigVM_Task, 'duration_secs': 0.306603} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.320077] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 3ea24d58-ffbd-41d4-a010-978a395ad2fc/3ea24d58-ffbd-41d4-a010-978a395ad2fc.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.321665] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-477254d2-d20b-4a9f-b8d7-c86cc3610650 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.329869] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 986.329869] env[61906]: value = "task-1357024" [ 986.329869] env[61906]: _type = "Task" [ 986.329869] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.339498] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357024, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.377427] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aafafb5d-3a20-4eb7-b4ae-5f430b17165e tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "cb879dc0-3af7-4279-aa28-66e2b8b4286d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.537s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.457248] env[61906]: DEBUG nova.compute.manager [req-f34550de-4931-4e50-8815-fbde7112a6b2 req-04353b2c-63f2-4991-82c2-c80555d59fc9 service nova] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Received event network-vif-deleted-e291dedf-5c3f-4d78-bad3-295b2d1e2f75 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.457248] env[61906]: INFO nova.compute.manager [req-f34550de-4931-4e50-8815-fbde7112a6b2 req-04353b2c-63f2-4991-82c2-c80555d59fc9 service nova] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Neutron deleted interface e291dedf-5c3f-4d78-bad3-295b2d1e2f75; detaching it from the instance and deleting it from the info cache [ 986.457248] env[61906]: DEBUG nova.network.neutron [req-f34550de-4931-4e50-8815-fbde7112a6b2 req-04353b2c-63f2-4991-82c2-c80555d59fc9 service nova] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.461311] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02949af6-d523-4fd2-b422-43e101451f2c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.470038] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f08f3ac-6762-4669-aa7c-bc6f340660ec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.500566] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982cd4b7-b66e-4d9f-9c76-4264ff5ab0a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.507948] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345829f7-59f7-4c09-89f6-9113c70b9451 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.521474] env[61906]: DEBUG nova.compute.provider_tree [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 986.577896] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.578105] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquired lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.578290] env[61906]: DEBUG nova.network.neutron [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.578480] env[61906]: DEBUG nova.objects.instance [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lazy-loading 'info_cache' on Instance uuid 082b1a00-dd07-4d37-860c-6a91d6d40eaf {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.841891] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357024, 'name': Rename_Task, 'duration_secs': 0.145286} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.842233] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 986.842527] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61b77fb7-816a-451c-882a-606ed7345c2d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.849012] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 986.849012] env[61906]: value = "task-1357025" [ 986.849012] env[61906]: _type = "Task" [ 986.849012] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.856330] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357025, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.928996] env[61906]: DEBUG nova.network.neutron [-] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.965021] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bd183ccc-a5a0-4377-8add-743015c051a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.974962] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f28c5d-09e8-4e6e-aa64-7793b1ceddcc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.000397] env[61906]: DEBUG nova.compute.manager [req-f34550de-4931-4e50-8815-fbde7112a6b2 req-04353b2c-63f2-4991-82c2-c80555d59fc9 service nova] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Detach interface failed, port_id=e291dedf-5c3f-4d78-bad3-295b2d1e2f75, reason: Instance 79452791-59cb-4722-bb4a-8e59d8c4e641 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 987.052030] env[61906]: ERROR nova.scheduler.client.report [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [req-2ecb2d18-bcb3-4ce8-aebb-89a0fc839f6e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6e524f5d-985e-41eb-a7ae-996be3ae1c1a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2ecb2d18-bcb3-4ce8-aebb-89a0fc839f6e"}]} [ 987.070210] env[61906]: DEBUG nova.scheduler.client.report [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Refreshing inventories for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 987.089106] env[61906]: DEBUG nova.scheduler.client.report [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Updating ProviderTree inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 987.089359] env[61906]: DEBUG nova.compute.provider_tree [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 987.102243] env[61906]: DEBUG nova.scheduler.client.report [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Refreshing aggregate associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 987.123261] env[61906]: DEBUG nova.scheduler.client.report [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Refreshing trait associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 987.219440] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-441c77f0-c941-4691-ada0-6bc33c80c34c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.227262] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a65b009-338f-4fcd-87b7-ea466cbfdc9e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.257524] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2065cf52-571b-4b18-a526-2afee988ac17 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.265661] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31097bfc-2a3e-4536-97a0-84ba84e3dd5a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.280765] env[61906]: DEBUG nova.compute.provider_tree [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 987.360674] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357025, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.433932] env[61906]: INFO nova.compute.manager [-] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Took 1.38 seconds to deallocate network for instance. [ 987.812793] env[61906]: ERROR nova.scheduler.client.report [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [req-f6520cb4-32c2-4f56-834f-bbba78051a55] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6e524f5d-985e-41eb-a7ae-996be3ae1c1a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n update conflict: Another thread concurrently updated the resource provider data. Please retry your update ", "code": "placement.concurrent_update", "request_id": "req-f6520cb4-32c2-4f56-834f-bbba78051a55"}]} [ 987.816882] env[61906]: DEBUG nova.network.neutron [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance_info_cache with network_info: [{"id": "0c3f7a1f-6d39-4063-b093-29db43118645", "address": "fa:16:3e:77:db:95", "network": {"id": "2b957ec8-0ca8-4ea0-acae-287f09c02dd9", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-250257405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3093466de87a45a2ac673c25bea19bc3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "30f1dacf-8988-41b8-aa8f-e9530f65ef46", "external-id": "nsx-vlan-transportzone-561", "segmentation_id": 561, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c3f7a1f-6d", "ovs_interfaceid": "0c3f7a1f-6d39-4063-b093-29db43118645", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.826717] env[61906]: DEBUG nova.scheduler.client.report [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Refreshing inventories for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 987.840725] env[61906]: DEBUG nova.scheduler.client.report [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Updating ProviderTree inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 987.840959] env[61906]: DEBUG nova.compute.provider_tree [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 987.854568] env[61906]: DEBUG nova.scheduler.client.report [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Refreshing aggregate associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 987.861984] env[61906]: DEBUG oslo_vmware.api [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357025, 'name': PowerOnVM_Task, 'duration_secs': 0.52794} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.862261] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 987.862465] env[61906]: INFO nova.compute.manager [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Took 7.59 seconds to spawn the instance on the hypervisor. [ 987.862645] env[61906]: DEBUG nova.compute.manager [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.863462] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52e325d-bfe5-4e89-96ce-7a337f9e25f6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.872364] env[61906]: DEBUG nova.scheduler.client.report [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Refreshing trait associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 987.940040] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.964300] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26807ef2-b57b-48d6-a834-d16fe8f80456 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.971892] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c74e74-7c95-4a2a-90dc-984d7626f3cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.000809] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27d800f-9c8b-4913-a811-0ee8c9577187 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.007526] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ab99e6-77b0-486d-8da9-213c87157cc9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.021479] env[61906]: DEBUG nova.compute.provider_tree [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 988.258765] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.259068] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.319647] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Releasing lock "refresh_cache-082b1a00-dd07-4d37-860c-6a91d6d40eaf" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.319647] env[61906]: DEBUG nova.objects.instance [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lazy-loading 'migration_context' on Instance uuid 082b1a00-dd07-4d37-860c-6a91d6d40eaf {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.384272] env[61906]: INFO nova.compute.manager [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Took 12.33 seconds to build instance. [ 988.549991] env[61906]: DEBUG nova.scheduler.client.report [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Updated inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with generation 113 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 988.550311] env[61906]: DEBUG nova.compute.provider_tree [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Updating resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a generation from 113 to 114 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 988.550506] env[61906]: DEBUG nova.compute.provider_tree [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 988.761826] env[61906]: DEBUG nova.compute.manager [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 988.822691] env[61906]: DEBUG nova.objects.base [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Object Instance<082b1a00-dd07-4d37-860c-6a91d6d40eaf> lazy-loaded attributes: info_cache,migration_context {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 988.823164] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22830829-60a4-4d7e-93ab-a62682a90c5d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.842747] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-021e546e-af01-4335-b37d-d253200bfb9f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.848147] env[61906]: DEBUG oslo_vmware.api [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 988.848147] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525047c9-53ce-2508-554c-16dde397d1d3" [ 988.848147] env[61906]: _type = "Task" [ 988.848147] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.856704] env[61906]: DEBUG oslo_vmware.api [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525047c9-53ce-2508-554c-16dde397d1d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.886357] env[61906]: DEBUG oslo_concurrency.lockutils [None req-30dc38d0-8eb8-4db5-86b3-673d11ee6f44 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.846s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.055133] env[61906]: DEBUG oslo_concurrency.lockutils [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.212s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.057517] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.118s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.057766] env[61906]: DEBUG nova.objects.instance [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lazy-loading 'resources' on Instance uuid 79452791-59cb-4722-bb4a-8e59d8c4e641 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.078655] env[61906]: INFO nova.scheduler.client.report [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleted allocations for instance e3578f5a-bd05-41fa-92e8-61d0e4ceed3f [ 989.280419] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.360839] env[61906]: DEBUG oslo_vmware.api [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525047c9-53ce-2508-554c-16dde397d1d3, 'name': SearchDatastore_Task, 'duration_secs': 0.007026} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.361167] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.587985] env[61906]: DEBUG oslo_concurrency.lockutils [None req-25c8e6c8-f7be-4824-bf01-56da78b745cb tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e3578f5a-bd05-41fa-92e8-61d0e4ceed3f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.182s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.653669] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4309aa1e-72ce-43d7-a4fe-f8b9e4d9afbe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.661532] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342a7b64-2e6b-40fe-b42b-c3ec1bda42d3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.691664] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549d51a8-bf2e-4ed0-822f-82073cd9b880 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.699194] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fce198f-e991-4c01-9888-9f9177ce66ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.712249] env[61906]: DEBUG nova.compute.provider_tree [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.756518] env[61906]: DEBUG nova.compute.manager [req-4e35f962-05ca-4a48-9993-6b6ad558934c req-f5e9fdfb-d58b-4081-9520-734ded8e158b service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Received event network-changed-fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.756768] env[61906]: DEBUG nova.compute.manager [req-4e35f962-05ca-4a48-9993-6b6ad558934c req-f5e9fdfb-d58b-4081-9520-734ded8e158b service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Refreshing instance network info cache due to event network-changed-fc9835ce-e30c-4b58-8804-1062f5d166de. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 989.757016] env[61906]: DEBUG oslo_concurrency.lockutils [req-4e35f962-05ca-4a48-9993-6b6ad558934c req-f5e9fdfb-d58b-4081-9520-734ded8e158b service nova] Acquiring lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.757179] env[61906]: DEBUG oslo_concurrency.lockutils [req-4e35f962-05ca-4a48-9993-6b6ad558934c req-f5e9fdfb-d58b-4081-9520-734ded8e158b service nova] Acquired lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.757345] env[61906]: DEBUG nova.network.neutron [req-4e35f962-05ca-4a48-9993-6b6ad558934c req-f5e9fdfb-d58b-4081-9520-734ded8e158b service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Refreshing network info cache for port fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 990.215053] env[61906]: DEBUG nova.scheduler.client.report [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.348035] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "729f5565-8cda-4bee-8644-fb94a06034fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.348348] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "729f5565-8cda-4bee-8644-fb94a06034fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.134304] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.077s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.136088] env[61906]: DEBUG nova.compute.manager [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 991.138730] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.858s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.140321] env[61906]: INFO nova.compute.claims [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 991.153565] env[61906]: INFO nova.scheduler.client.report [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Deleted allocations for instance 79452791-59cb-4722-bb4a-8e59d8c4e641 [ 991.237067] env[61906]: DEBUG nova.network.neutron [req-4e35f962-05ca-4a48-9993-6b6ad558934c req-f5e9fdfb-d58b-4081-9520-734ded8e158b service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updated VIF entry in instance network info cache for port fc9835ce-e30c-4b58-8804-1062f5d166de. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.237490] env[61906]: DEBUG nova.network.neutron [req-4e35f962-05ca-4a48-9993-6b6ad558934c req-f5e9fdfb-d58b-4081-9520-734ded8e158b service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updating instance_info_cache with network_info: [{"id": "fc9835ce-e30c-4b58-8804-1062f5d166de", "address": "fa:16:3e:b8:59:0b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc9835ce-e3", "ovs_interfaceid": "fc9835ce-e30c-4b58-8804-1062f5d166de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.661964] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d47fc72e-5a3d-494d-9b8c-3644c4ef003d tempest-ServerRescueNegativeTestJSON-934988020 tempest-ServerRescueNegativeTestJSON-934988020-project-member] Lock "79452791-59cb-4722-bb4a-8e59d8c4e641" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.733s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.663930] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.739848] env[61906]: DEBUG oslo_concurrency.lockutils [req-4e35f962-05ca-4a48-9993-6b6ad558934c req-f5e9fdfb-d58b-4081-9520-734ded8e158b service nova] Releasing lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.791780] env[61906]: DEBUG nova.compute.manager [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Stashing vm_state: active {{(pid=61906) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 992.243115] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d6720f8-f0d6-4357-b277-1405a75b5d66 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.251864] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f96a5b7-6661-4fb5-9a67-68d6f77f15ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.282112] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a2899f-eb5c-4353-95c8-6196550adf73 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.289474] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29613dd8-251f-47f8-9499-6ecfe3cfaffc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.305022] env[61906]: DEBUG nova.compute.provider_tree [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.316341] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.811000] env[61906]: DEBUG nova.scheduler.client.report [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.317642] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.179s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.318271] env[61906]: DEBUG nova.compute.manager [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 993.321405] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.960s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.824854] env[61906]: DEBUG nova.compute.utils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 993.828884] env[61906]: DEBUG nova.compute.manager [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 993.829121] env[61906]: DEBUG nova.network.neutron [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 993.870368] env[61906]: DEBUG nova.policy [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f084b58b30a64e06ad69e8f6b366207c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f3bb9ce83b3430bb9202d3c72e77b4c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 993.920418] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f3a094-82b0-41fb-bbe5-1836a8ea6446 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.928252] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8781b19b-3a2b-423b-9942-362bbe94a89f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.960464] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4346dd2-9a60-4e06-ab95-20ccc0a0fba1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.969047] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eff7796e-328c-476a-8a71-1d3a3515d3a1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.983762] env[61906]: DEBUG nova.compute.provider_tree [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.133403] env[61906]: DEBUG nova.network.neutron [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Successfully created port: 3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 994.329461] env[61906]: DEBUG nova.compute.manager [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 994.487399] env[61906]: DEBUG nova.scheduler.client.report [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 995.339523] env[61906]: DEBUG nova.compute.manager [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 995.363564] env[61906]: DEBUG nova.virt.hardware [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 995.363830] env[61906]: DEBUG nova.virt.hardware [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 995.363979] env[61906]: DEBUG nova.virt.hardware [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 995.364183] env[61906]: DEBUG nova.virt.hardware [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 995.364337] env[61906]: DEBUG nova.virt.hardware [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 995.364841] env[61906]: DEBUG nova.virt.hardware [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 995.364841] env[61906]: DEBUG nova.virt.hardware [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 995.364961] env[61906]: DEBUG nova.virt.hardware [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 995.365042] env[61906]: DEBUG nova.virt.hardware [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 995.365206] env[61906]: DEBUG nova.virt.hardware [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 995.365384] env[61906]: DEBUG nova.virt.hardware [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.366248] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5719600-217f-4333-8b1b-68db8cce4abb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.374220] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f455b09-9413-4a1c-9297-765802952994 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.497873] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.176s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.500875] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.837s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.502378] env[61906]: INFO nova.compute.claims [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 995.519034] env[61906]: DEBUG nova.compute.manager [req-78086eca-43dd-4912-97bd-a2da2dc864f0 req-85b29ad2-63cc-47d3-8ba0-3d11221c92e2 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Received event network-vif-plugged-3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.520286] env[61906]: DEBUG oslo_concurrency.lockutils [req-78086eca-43dd-4912-97bd-a2da2dc864f0 req-85b29ad2-63cc-47d3-8ba0-3d11221c92e2 service nova] Acquiring lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.520286] env[61906]: DEBUG oslo_concurrency.lockutils [req-78086eca-43dd-4912-97bd-a2da2dc864f0 req-85b29ad2-63cc-47d3-8ba0-3d11221c92e2 service nova] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.520286] env[61906]: DEBUG oslo_concurrency.lockutils [req-78086eca-43dd-4912-97bd-a2da2dc864f0 req-85b29ad2-63cc-47d3-8ba0-3d11221c92e2 service nova] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.520286] env[61906]: DEBUG nova.compute.manager [req-78086eca-43dd-4912-97bd-a2da2dc864f0 req-85b29ad2-63cc-47d3-8ba0-3d11221c92e2 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] No waiting events found dispatching network-vif-plugged-3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 995.520488] env[61906]: WARNING nova.compute.manager [req-78086eca-43dd-4912-97bd-a2da2dc864f0 req-85b29ad2-63cc-47d3-8ba0-3d11221c92e2 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Received unexpected event network-vif-plugged-3c075a5e-3c44-45b5-bada-620921c0a005 for instance with vm_state building and task_state spawning. [ 995.609282] env[61906]: DEBUG nova.network.neutron [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Successfully updated port: 3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 996.060739] env[61906]: INFO nova.scheduler.client.report [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleted allocation for migration 64b6b34a-47b3-4cdf-b34b-951bf173cffb [ 996.111151] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.111308] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.111456] env[61906]: DEBUG nova.network.neutron [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 996.566508] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.528s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.592556] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357e679c-815d-4c38-b5fc-cbb4bf581eaa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.600949] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b63ad9f-21bd-44a0-8640-a9e2506b0721 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.631464] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2730874-9742-466b-9311-ca223ddd8fb1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.638922] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67df3c1f-8ac8-47ad-9ddb-6e0e39738ebf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.651822] env[61906]: DEBUG nova.compute.provider_tree [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.660942] env[61906]: DEBUG nova.network.neutron [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 996.790131] env[61906]: DEBUG nova.network.neutron [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updating instance_info_cache with network_info: [{"id": "3c075a5e-3c44-45b5-bada-620921c0a005", "address": "fa:16:3e:39:60:50", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c075a5e-3c", "ovs_interfaceid": "3c075a5e-3c44-45b5-bada-620921c0a005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.155421] env[61906]: DEBUG nova.scheduler.client.report [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 997.292726] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.293102] env[61906]: DEBUG nova.compute.manager [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Instance network_info: |[{"id": "3c075a5e-3c44-45b5-bada-620921c0a005", "address": "fa:16:3e:39:60:50", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c075a5e-3c", "ovs_interfaceid": "3c075a5e-3c44-45b5-bada-620921c0a005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 997.293517] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:60:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8edfde4-5a99-4745-956d-04da82ab1b85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c075a5e-3c44-45b5-bada-620921c0a005', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 997.300825] env[61906]: DEBUG oslo.service.loopingcall [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.301454] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 997.301716] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-612aae21-5dac-4378-aa4d-15dc9b75aa35 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.320830] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 997.320830] env[61906]: value = "task-1357026" [ 997.320830] env[61906]: _type = "Task" [ 997.320830] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.328194] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357026, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.588923] env[61906]: DEBUG nova.compute.manager [req-015db120-3d93-42ea-8a8b-a17b19a220eb req-e47835f8-c4c2-4d3f-a7f0-b3dc4dd8e91b service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Received event network-changed-3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.589360] env[61906]: DEBUG nova.compute.manager [req-015db120-3d93-42ea-8a8b-a17b19a220eb req-e47835f8-c4c2-4d3f-a7f0-b3dc4dd8e91b service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Refreshing instance network info cache due to event network-changed-3c075a5e-3c44-45b5-bada-620921c0a005. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 997.589437] env[61906]: DEBUG oslo_concurrency.lockutils [req-015db120-3d93-42ea-8a8b-a17b19a220eb req-e47835f8-c4c2-4d3f-a7f0-b3dc4dd8e91b service nova] Acquiring lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.589676] env[61906]: DEBUG oslo_concurrency.lockutils [req-015db120-3d93-42ea-8a8b-a17b19a220eb req-e47835f8-c4c2-4d3f-a7f0-b3dc4dd8e91b service nova] Acquired lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.589773] env[61906]: DEBUG nova.network.neutron [req-015db120-3d93-42ea-8a8b-a17b19a220eb req-e47835f8-c4c2-4d3f-a7f0-b3dc4dd8e91b service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Refreshing network info cache for port 3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 997.660104] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.159s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.660624] env[61906]: DEBUG nova.compute.manager [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 997.663282] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.347s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.831134] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357026, 'name': CreateVM_Task, 'duration_secs': 0.49873} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.831362] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 997.831996] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.832198] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.832527] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 997.832783] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c967947a-b178-43ab-9230-b53f090c8334 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.837400] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 997.837400] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ed3ed1-ae25-be00-16d6-442f477566e6" [ 997.837400] env[61906]: _type = "Task" [ 997.837400] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.844470] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ed3ed1-ae25-be00-16d6-442f477566e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.066479] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.066767] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.066996] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.067229] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.067411] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.069545] env[61906]: INFO nova.compute.manager [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Terminating instance [ 998.071339] env[61906]: DEBUG nova.compute.manager [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 998.071542] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 998.072376] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1ace72-f5d3-4dc5-a253-eea27f16bbe4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.079937] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 998.080205] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7823ca93-ff83-4639-ad8f-b610838fc7fd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.085931] env[61906]: DEBUG oslo_vmware.api [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 998.085931] env[61906]: value = "task-1357027" [ 998.085931] env[61906]: _type = "Task" [ 998.085931] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.095170] env[61906]: DEBUG oslo_vmware.api [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1357027, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.166286] env[61906]: DEBUG nova.compute.utils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 998.169955] env[61906]: INFO nova.compute.claims [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 998.175529] env[61906]: DEBUG nova.compute.manager [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 998.175707] env[61906]: DEBUG nova.network.neutron [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 998.214506] env[61906]: DEBUG nova.policy [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73219f5e96f64d1da86a19da9b18b06a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5533be74874b4094b7ef0f98a121f60b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 998.293817] env[61906]: DEBUG nova.network.neutron [req-015db120-3d93-42ea-8a8b-a17b19a220eb req-e47835f8-c4c2-4d3f-a7f0-b3dc4dd8e91b service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updated VIF entry in instance network info cache for port 3c075a5e-3c44-45b5-bada-620921c0a005. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 998.294234] env[61906]: DEBUG nova.network.neutron [req-015db120-3d93-42ea-8a8b-a17b19a220eb req-e47835f8-c4c2-4d3f-a7f0-b3dc4dd8e91b service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updating instance_info_cache with network_info: [{"id": "3c075a5e-3c44-45b5-bada-620921c0a005", "address": "fa:16:3e:39:60:50", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c075a5e-3c", "ovs_interfaceid": "3c075a5e-3c44-45b5-bada-620921c0a005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.349675] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ed3ed1-ae25-be00-16d6-442f477566e6, 'name': SearchDatastore_Task, 'duration_secs': 0.009889} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.349976] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.350276] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 998.350529] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.350682] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.350867] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 998.351143] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-36896382-0056-4c51-870e-0a04a6046826 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.358403] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 998.358562] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 998.359264] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-629d3f7a-37fa-43ca-96a5-e0939d9e99f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.363676] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 998.363676] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525b2ed1-c651-4add-af32-c577dc46ee38" [ 998.363676] env[61906]: _type = "Task" [ 998.363676] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.371033] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525b2ed1-c651-4add-af32-c577dc46ee38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.529068] env[61906]: DEBUG nova.network.neutron [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Successfully created port: a6bc2330-c2b4-4d0b-8178-f827316508a3 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 998.596281] env[61906]: DEBUG oslo_vmware.api [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1357027, 'name': PowerOffVM_Task, 'duration_secs': 0.177848} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.596703] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 998.596703] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 998.596939] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-42bec8be-6160-4be3-aaf5-344e87f0267a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.658786] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 998.658985] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 998.659156] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleting the datastore file [datastore2] 082b1a00-dd07-4d37-860c-6a91d6d40eaf {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.659394] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f384bb60-b5ee-4a15-873e-7e3812ce0816 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.665301] env[61906]: DEBUG oslo_vmware.api [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for the task: (returnval){ [ 998.665301] env[61906]: value = "task-1357029" [ 998.665301] env[61906]: _type = "Task" [ 998.665301] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.672976] env[61906]: DEBUG oslo_vmware.api [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1357029, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.674567] env[61906]: DEBUG nova.compute.manager [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 998.679076] env[61906]: INFO nova.compute.resource_tracker [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating resource usage from migration f9f0d985-c6eb-4adc-adab-c9f39bfc4d5c [ 998.787387] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d180490d-fb18-4987-8653-f3bc115b3b20 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.795658] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd2c46c-e42a-4572-9234-e6f3bf3a16e1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.799664] env[61906]: DEBUG oslo_concurrency.lockutils [req-015db120-3d93-42ea-8a8b-a17b19a220eb req-e47835f8-c4c2-4d3f-a7f0-b3dc4dd8e91b service nova] Releasing lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.827051] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e748b2bf-f44c-4e00-acec-026093d9a597 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.834107] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31d0534-deec-4b0d-84b3-e8c194d0e5df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.846907] env[61906]: DEBUG nova.compute.provider_tree [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.873747] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525b2ed1-c651-4add-af32-c577dc46ee38, 'name': SearchDatastore_Task, 'duration_secs': 0.008072} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.874488] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4bcc8f4c-9d9f-47b3-87e6-7282114e0131 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.878921] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 998.878921] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52627a29-baea-4fa0-e42f-8f70648e1ec1" [ 998.878921] env[61906]: _type = "Task" [ 998.878921] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.886921] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52627a29-baea-4fa0-e42f-8f70648e1ec1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.175345] env[61906]: DEBUG oslo_vmware.api [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Task: {'id': task-1357029, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139927} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.175663] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.175868] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 999.176062] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 999.176253] env[61906]: INFO nova.compute.manager [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Took 1.10 seconds to destroy the instance on the hypervisor. [ 999.176485] env[61906]: DEBUG oslo.service.loopingcall [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.176681] env[61906]: DEBUG nova.compute.manager [-] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 999.176773] env[61906]: DEBUG nova.network.neutron [-] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 999.349866] env[61906]: DEBUG nova.scheduler.client.report [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 999.390173] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52627a29-baea-4fa0-e42f-8f70648e1ec1, 'name': SearchDatastore_Task, 'duration_secs': 0.009437} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.390443] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.390704] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] d8c1fc04-d5ac-4c78-95e1-e135b22819c0/d8c1fc04-d5ac-4c78-95e1-e135b22819c0.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 999.390952] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d5807c48-18df-43a2-8e20-093471855e24 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.398223] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 999.398223] env[61906]: value = "task-1357030" [ 999.398223] env[61906]: _type = "Task" [ 999.398223] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.406055] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357030, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.668542] env[61906]: DEBUG nova.compute.manager [req-774310b9-a68c-453a-9482-aff5bf9f1920 req-8a218263-f526-4675-be07-9210744f50b1 service nova] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Received event network-vif-deleted-0c3f7a1f-6d39-4063-b093-29db43118645 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 999.668824] env[61906]: INFO nova.compute.manager [req-774310b9-a68c-453a-9482-aff5bf9f1920 req-8a218263-f526-4675-be07-9210744f50b1 service nova] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Neutron deleted interface 0c3f7a1f-6d39-4063-b093-29db43118645; detaching it from the instance and deleting it from the info cache [ 999.668958] env[61906]: DEBUG nova.network.neutron [req-774310b9-a68c-453a-9482-aff5bf9f1920 req-8a218263-f526-4675-be07-9210744f50b1 service nova] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.684132] env[61906]: DEBUG nova.compute.manager [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 999.714965] env[61906]: DEBUG nova.virt.hardware [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 999.715269] env[61906]: DEBUG nova.virt.hardware [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 999.715471] env[61906]: DEBUG nova.virt.hardware [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 999.715691] env[61906]: DEBUG nova.virt.hardware [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 999.715860] env[61906]: DEBUG nova.virt.hardware [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 999.716034] env[61906]: DEBUG nova.virt.hardware [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 999.716265] env[61906]: DEBUG nova.virt.hardware [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 999.716531] env[61906]: DEBUG nova.virt.hardware [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 999.716739] env[61906]: DEBUG nova.virt.hardware [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 999.716924] env[61906]: DEBUG nova.virt.hardware [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 999.717176] env[61906]: DEBUG nova.virt.hardware [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 999.718509] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b035f59-a077-4045-807b-34d1106c69ef {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.727743] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb6b706-a396-4720-9e22-1de308aedd30 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.854479] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.191s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.854747] env[61906]: INFO nova.compute.manager [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Migrating [ 999.878849] env[61906]: DEBUG nova.network.neutron [-] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.908061] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357030, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479714} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.908339] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] d8c1fc04-d5ac-4c78-95e1-e135b22819c0/d8c1fc04-d5ac-4c78-95e1-e135b22819c0.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 999.908555] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 999.908792] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b014aa48-666b-4d14-aa2a-66370266c264 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.915076] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 999.915076] env[61906]: value = "task-1357031" [ 999.915076] env[61906]: _type = "Task" [ 999.915076] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.923370] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357031, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.006367] env[61906]: DEBUG nova.compute.manager [req-95b55d46-a9d1-482c-bc52-9fc7ce363b3c req-995fb4c2-dd64-4f58-8096-ae3d331946d9 service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Received event network-vif-plugged-a6bc2330-c2b4-4d0b-8178-f827316508a3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.006593] env[61906]: DEBUG oslo_concurrency.lockutils [req-95b55d46-a9d1-482c-bc52-9fc7ce363b3c req-995fb4c2-dd64-4f58-8096-ae3d331946d9 service nova] Acquiring lock "729f5565-8cda-4bee-8644-fb94a06034fb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.006812] env[61906]: DEBUG oslo_concurrency.lockutils [req-95b55d46-a9d1-482c-bc52-9fc7ce363b3c req-995fb4c2-dd64-4f58-8096-ae3d331946d9 service nova] Lock "729f5565-8cda-4bee-8644-fb94a06034fb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.006979] env[61906]: DEBUG oslo_concurrency.lockutils [req-95b55d46-a9d1-482c-bc52-9fc7ce363b3c req-995fb4c2-dd64-4f58-8096-ae3d331946d9 service nova] Lock "729f5565-8cda-4bee-8644-fb94a06034fb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.007217] env[61906]: DEBUG nova.compute.manager [req-95b55d46-a9d1-482c-bc52-9fc7ce363b3c req-995fb4c2-dd64-4f58-8096-ae3d331946d9 service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] No waiting events found dispatching network-vif-plugged-a6bc2330-c2b4-4d0b-8178-f827316508a3 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1000.007407] env[61906]: WARNING nova.compute.manager [req-95b55d46-a9d1-482c-bc52-9fc7ce363b3c req-995fb4c2-dd64-4f58-8096-ae3d331946d9 service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Received unexpected event network-vif-plugged-a6bc2330-c2b4-4d0b-8178-f827316508a3 for instance with vm_state building and task_state spawning. [ 1000.085504] env[61906]: DEBUG nova.network.neutron [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Successfully updated port: a6bc2330-c2b4-4d0b-8178-f827316508a3 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1000.172132] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5c9c912-e93a-4ba9-9917-12b0bbc3426b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.181484] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06c04cb-5077-492f-948e-795b9d5bbcda {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.204411] env[61906]: DEBUG nova.compute.manager [req-774310b9-a68c-453a-9482-aff5bf9f1920 req-8a218263-f526-4675-be07-9210744f50b1 service nova] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Detach interface failed, port_id=0c3f7a1f-6d39-4063-b093-29db43118645, reason: Instance 082b1a00-dd07-4d37-860c-6a91d6d40eaf could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1000.370074] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.370207] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.370382] env[61906]: DEBUG nova.network.neutron [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1000.382671] env[61906]: INFO nova.compute.manager [-] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Took 1.21 seconds to deallocate network for instance. [ 1000.425468] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357031, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062968} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.426323] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1000.427108] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b94ab56-66fa-43f5-9e99-64ec6a1c71da {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.448869] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] d8c1fc04-d5ac-4c78-95e1-e135b22819c0/d8c1fc04-d5ac-4c78-95e1-e135b22819c0.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.449383] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8ddf19c-1ba3-44f2-8d85-414a39cd0839 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.468240] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1000.468240] env[61906]: value = "task-1357032" [ 1000.468240] env[61906]: _type = "Task" [ 1000.468240] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.475992] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357032, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.589773] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.589967] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.590122] env[61906]: DEBUG nova.network.neutron [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1000.888975] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.889432] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.889750] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.925407] env[61906]: INFO nova.scheduler.client.report [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Deleted allocations for instance 082b1a00-dd07-4d37-860c-6a91d6d40eaf [ 1000.981062] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357032, 'name': ReconfigVM_Task, 'duration_secs': 0.316731} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.981622] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Reconfigured VM instance instance-00000061 to attach disk [datastore2] d8c1fc04-d5ac-4c78-95e1-e135b22819c0/d8c1fc04-d5ac-4c78-95e1-e135b22819c0.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.981955] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa6e8977-01a8-4ca1-8f10-253c177a738a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.988405] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1000.988405] env[61906]: value = "task-1357033" [ 1000.988405] env[61906]: _type = "Task" [ 1000.988405] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.999193] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357033, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.104817] env[61906]: DEBUG nova.network.neutron [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance_info_cache with network_info: [{"id": "d803c8e8-7145-412c-aa3e-2545666742eb", "address": "fa:16:3e:46:59:77", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd803c8e8-71", "ovs_interfaceid": "d803c8e8-7145-412c-aa3e-2545666742eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.120264] env[61906]: DEBUG nova.network.neutron [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1001.239752] env[61906]: DEBUG nova.network.neutron [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updating instance_info_cache with network_info: [{"id": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "address": "fa:16:3e:87:df:8b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6bc2330-c2", "ovs_interfaceid": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.434964] env[61906]: DEBUG oslo_concurrency.lockutils [None req-23f7096a-bdb4-41f3-b78c-44617ab041a7 tempest-DeleteServersTestJSON-874841312 tempest-DeleteServersTestJSON-874841312-project-member] Lock "082b1a00-dd07-4d37-860c-6a91d6d40eaf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.368s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.498300] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357033, 'name': Rename_Task, 'duration_secs': 0.135441} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.498582] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1001.498826] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-039ba6d2-ca07-4970-9284-b4b25c31d1ca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.504751] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1001.504751] env[61906]: value = "task-1357034" [ 1001.504751] env[61906]: _type = "Task" [ 1001.504751] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.511764] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357034, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.608630] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.742123] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.742428] env[61906]: DEBUG nova.compute.manager [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Instance network_info: |[{"id": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "address": "fa:16:3e:87:df:8b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6bc2330-c2", "ovs_interfaceid": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1001.742861] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:df:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a6bc2330-c2b4-4d0b-8178-f827316508a3', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1001.751048] env[61906]: DEBUG oslo.service.loopingcall [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.751261] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1001.751489] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7785cf74-ff9a-426f-8dba-ec2914b9f335 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.771573] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1001.771573] env[61906]: value = "task-1357035" [ 1001.771573] env[61906]: _type = "Task" [ 1001.771573] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.778994] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357035, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.014773] env[61906]: DEBUG oslo_vmware.api [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357034, 'name': PowerOnVM_Task, 'duration_secs': 0.450146} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.015083] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1002.015295] env[61906]: INFO nova.compute.manager [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Took 6.68 seconds to spawn the instance on the hypervisor. [ 1002.015512] env[61906]: DEBUG nova.compute.manager [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.016313] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07408e3b-04bc-4f68-ab9b-0c38db50964b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.030443] env[61906]: DEBUG nova.compute.manager [req-acc5a09b-102d-4788-8ed2-61b1706a2774 req-73d07956-70c5-4dfb-b37e-137aaced77ae service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Received event network-changed-a6bc2330-c2b4-4d0b-8178-f827316508a3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.030610] env[61906]: DEBUG nova.compute.manager [req-acc5a09b-102d-4788-8ed2-61b1706a2774 req-73d07956-70c5-4dfb-b37e-137aaced77ae service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Refreshing instance network info cache due to event network-changed-a6bc2330-c2b4-4d0b-8178-f827316508a3. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1002.030837] env[61906]: DEBUG oslo_concurrency.lockutils [req-acc5a09b-102d-4788-8ed2-61b1706a2774 req-73d07956-70c5-4dfb-b37e-137aaced77ae service nova] Acquiring lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.030986] env[61906]: DEBUG oslo_concurrency.lockutils [req-acc5a09b-102d-4788-8ed2-61b1706a2774 req-73d07956-70c5-4dfb-b37e-137aaced77ae service nova] Acquired lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.031357] env[61906]: DEBUG nova.network.neutron [req-acc5a09b-102d-4788-8ed2-61b1706a2774 req-73d07956-70c5-4dfb-b37e-137aaced77ae service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Refreshing network info cache for port a6bc2330-c2b4-4d0b-8178-f827316508a3 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1002.283034] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357035, 'name': CreateVM_Task, 'duration_secs': 0.466776} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.283034] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1002.283797] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.284022] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.284401] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1002.284698] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc8a7327-0613-46e3-8b71-acc5b7e5051c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.289909] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1002.289909] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52dcb8fe-c7bf-4e04-1df8-94e0a4ddbec3" [ 1002.289909] env[61906]: _type = "Task" [ 1002.289909] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.301542] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52dcb8fe-c7bf-4e04-1df8-94e0a4ddbec3, 'name': SearchDatastore_Task, 'duration_secs': 0.009045} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.301890] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.302172] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1002.302418] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.302585] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.302821] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1002.303126] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3b76eb2a-80de-4ad7-80db-560ed76dcdb6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.311193] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1002.311420] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1002.312239] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1465ce38-cd01-4aa9-8104-6e7445c15a9a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.317638] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1002.317638] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52e73e85-3f71-87d0-7573-7fca80fc7c05" [ 1002.317638] env[61906]: _type = "Task" [ 1002.317638] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.325188] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52e73e85-3f71-87d0-7573-7fca80fc7c05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.538844] env[61906]: INFO nova.compute.manager [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Took 13.27 seconds to build instance. [ 1002.729651] env[61906]: DEBUG nova.network.neutron [req-acc5a09b-102d-4788-8ed2-61b1706a2774 req-73d07956-70c5-4dfb-b37e-137aaced77ae service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updated VIF entry in instance network info cache for port a6bc2330-c2b4-4d0b-8178-f827316508a3. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1002.730052] env[61906]: DEBUG nova.network.neutron [req-acc5a09b-102d-4788-8ed2-61b1706a2774 req-73d07956-70c5-4dfb-b37e-137aaced77ae service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updating instance_info_cache with network_info: [{"id": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "address": "fa:16:3e:87:df:8b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6bc2330-c2", "ovs_interfaceid": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.828041] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52e73e85-3f71-87d0-7573-7fca80fc7c05, 'name': SearchDatastore_Task, 'duration_secs': 0.009345} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.828806] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f526b063-7b3e-4ce8-8439-9750ae5e257d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.833606] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1002.833606] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b77a62-5b62-f286-84b8-4b2bf4c5fa8d" [ 1002.833606] env[61906]: _type = "Task" [ 1002.833606] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.840941] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b77a62-5b62-f286-84b8-4b2bf4c5fa8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.040818] env[61906]: DEBUG oslo_concurrency.lockutils [None req-21529044-8050-47cd-bfec-cfb380b999b9 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.782s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.123407] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1fa666-cb9c-42c1-8986-f9bc53188941 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.142633] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance 'e133605d-c630-4b9e-a314-bf496c853710' progress to 0 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1003.233198] env[61906]: DEBUG oslo_concurrency.lockutils [req-acc5a09b-102d-4788-8ed2-61b1706a2774 req-73d07956-70c5-4dfb-b37e-137aaced77ae service nova] Releasing lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.345250] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b77a62-5b62-f286-84b8-4b2bf4c5fa8d, 'name': SearchDatastore_Task, 'duration_secs': 0.008802} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.345507] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.345742] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 729f5565-8cda-4bee-8644-fb94a06034fb/729f5565-8cda-4bee-8644-fb94a06034fb.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1003.345997] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1fab1177-39b3-4863-a0de-27461cd10171 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.352970] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1003.352970] env[61906]: value = "task-1357037" [ 1003.352970] env[61906]: _type = "Task" [ 1003.352970] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.360741] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357037, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.650289] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.650759] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-006cd450-5619-4f7e-a3da-ea355a68793b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.658430] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1003.658430] env[61906]: value = "task-1357038" [ 1003.658430] env[61906]: _type = "Task" [ 1003.658430] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.668363] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357038, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.863251] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357037, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.106267] env[61906]: DEBUG nova.compute.manager [req-e7f27caa-57c4-43ea-b657-9177cbbf2b3e req-55ad023e-07a5-4d0d-b4a9-32edb5ab2f7f service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Received event network-changed-3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.106267] env[61906]: DEBUG nova.compute.manager [req-e7f27caa-57c4-43ea-b657-9177cbbf2b3e req-55ad023e-07a5-4d0d-b4a9-32edb5ab2f7f service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Refreshing instance network info cache due to event network-changed-3c075a5e-3c44-45b5-bada-620921c0a005. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1004.106267] env[61906]: DEBUG oslo_concurrency.lockutils [req-e7f27caa-57c4-43ea-b657-9177cbbf2b3e req-55ad023e-07a5-4d0d-b4a9-32edb5ab2f7f service nova] Acquiring lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.106650] env[61906]: DEBUG oslo_concurrency.lockutils [req-e7f27caa-57c4-43ea-b657-9177cbbf2b3e req-55ad023e-07a5-4d0d-b4a9-32edb5ab2f7f service nova] Acquired lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.106650] env[61906]: DEBUG nova.network.neutron [req-e7f27caa-57c4-43ea-b657-9177cbbf2b3e req-55ad023e-07a5-4d0d-b4a9-32edb5ab2f7f service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Refreshing network info cache for port 3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1004.167832] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357038, 'name': PowerOffVM_Task, 'duration_secs': 0.19805} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.168072] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.168274] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance 'e133605d-c630-4b9e-a314-bf496c853710' progress to 17 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1004.363883] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357037, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550605} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.364191] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 729f5565-8cda-4bee-8644-fb94a06034fb/729f5565-8cda-4bee-8644-fb94a06034fb.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1004.364446] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1004.364721] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8479a933-76b2-4bd4-9aa1-4d979644bff6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.371222] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1004.371222] env[61906]: value = "task-1357039" [ 1004.371222] env[61906]: _type = "Task" [ 1004.371222] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.379509] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357039, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.674669] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1004.674953] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1004.675141] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1004.675384] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1004.675579] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1004.675736] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1004.675949] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1004.676237] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1004.676435] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1004.676623] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1004.676835] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1004.682182] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-073b7df2-6715-4552-94b3-3afcad4a9651 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.699486] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1004.699486] env[61906]: value = "task-1357040" [ 1004.699486] env[61906]: _type = "Task" [ 1004.699486] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.707822] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357040, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.880852] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357039, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.121319} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.881149] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.882834] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ecb5b5e-0047-4611-bb76-5f34a1c0fbe2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.907768] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 729f5565-8cda-4bee-8644-fb94a06034fb/729f5565-8cda-4bee-8644-fb94a06034fb.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.908135] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6548051-5613-496f-a229-02c70dddf569 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.929468] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1004.929468] env[61906]: value = "task-1357041" [ 1004.929468] env[61906]: _type = "Task" [ 1004.929468] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.943652] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357041, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.123950] env[61906]: DEBUG nova.network.neutron [req-e7f27caa-57c4-43ea-b657-9177cbbf2b3e req-55ad023e-07a5-4d0d-b4a9-32edb5ab2f7f service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updated VIF entry in instance network info cache for port 3c075a5e-3c44-45b5-bada-620921c0a005. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1005.124331] env[61906]: DEBUG nova.network.neutron [req-e7f27caa-57c4-43ea-b657-9177cbbf2b3e req-55ad023e-07a5-4d0d-b4a9-32edb5ab2f7f service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updating instance_info_cache with network_info: [{"id": "3c075a5e-3c44-45b5-bada-620921c0a005", "address": "fa:16:3e:39:60:50", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c075a5e-3c", "ovs_interfaceid": "3c075a5e-3c44-45b5-bada-620921c0a005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.209705] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357040, 'name': ReconfigVM_Task, 'duration_secs': 0.258224} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.210110] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance 'e133605d-c630-4b9e-a314-bf496c853710' progress to 33 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1005.440068] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.632184] env[61906]: DEBUG oslo_concurrency.lockutils [req-e7f27caa-57c4-43ea-b657-9177cbbf2b3e req-55ad023e-07a5-4d0d-b4a9-32edb5ab2f7f service nova] Releasing lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.717063] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.717431] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.717613] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.717823] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.717977] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.718193] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.718444] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.718620] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.718797] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.718966] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.719173] env[61906]: DEBUG nova.virt.hardware [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.724551] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Reconfiguring VM instance instance-00000046 to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1005.724844] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7fabe6ae-a20a-419f-9056-c04c57f2ebc0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.745626] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1005.745626] env[61906]: value = "task-1357042" [ 1005.745626] env[61906]: _type = "Task" [ 1005.745626] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.754043] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357042, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.943329] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.255531] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357042, 'name': ReconfigVM_Task, 'duration_secs': 0.296836} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.255842] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Reconfigured VM instance instance-00000046 to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1006.256639] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3460aca-d6eb-4950-8542-a9132a3e717a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.279493] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] e133605d-c630-4b9e-a314-bf496c853710/e133605d-c630-4b9e-a314-bf496c853710.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1006.280025] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77d0d350-e23a-4367-9304-6fbe4bdd8133 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.299856] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1006.299856] env[61906]: value = "task-1357043" [ 1006.299856] env[61906]: _type = "Task" [ 1006.299856] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.307893] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357043, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.440080] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357041, 'name': ReconfigVM_Task, 'duration_secs': 1.212554} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.440382] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 729f5565-8cda-4bee-8644-fb94a06034fb/729f5565-8cda-4bee-8644-fb94a06034fb.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.441011] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cc4c5b12-fc03-42c6-bc7c-d5ac583f9e2e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.446398] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1006.446398] env[61906]: value = "task-1357044" [ 1006.446398] env[61906]: _type = "Task" [ 1006.446398] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.453627] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357044, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.808967] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357043, 'name': ReconfigVM_Task, 'duration_secs': 0.247378} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.809533] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Reconfigured VM instance instance-00000046 to attach disk [datastore1] e133605d-c630-4b9e-a314-bf496c853710/e133605d-c630-4b9e-a314-bf496c853710.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.809645] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance 'e133605d-c630-4b9e-a314-bf496c853710' progress to 50 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1006.956415] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357044, 'name': Rename_Task, 'duration_secs': 0.129494} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.956706] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1006.956958] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb0dcc18-67e5-4504-ba9f-a4a5e2b2a0a3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.963479] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1006.963479] env[61906]: value = "task-1357045" [ 1006.963479] env[61906]: _type = "Task" [ 1006.963479] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.973710] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357045, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.316145] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591a02cd-e380-47ed-9a81-199d9536ea17 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.336206] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e510bda9-dac7-4c12-a20c-b0b6be9b83c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.355697] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance 'e133605d-c630-4b9e-a314-bf496c853710' progress to 67 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1007.475018] env[61906]: DEBUG oslo_vmware.api [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357045, 'name': PowerOnVM_Task, 'duration_secs': 0.474588} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.475018] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1007.475212] env[61906]: INFO nova.compute.manager [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Took 7.79 seconds to spawn the instance on the hypervisor. [ 1007.475389] env[61906]: DEBUG nova.compute.manager [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1007.476153] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700f0068-9e8f-472f-9ee2-97115947028f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.918064] env[61906]: DEBUG nova.network.neutron [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Port d803c8e8-7145-412c-aa3e-2545666742eb binding to destination host cpu-1 is already ACTIVE {{(pid=61906) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1007.994909] env[61906]: INFO nova.compute.manager [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Took 16.35 seconds to build instance. [ 1008.497681] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c662749-2a96-4134-a686-b32fda6b6134 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "729f5565-8cda-4bee-8644-fb94a06034fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.149s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.945160] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "e133605d-c630-4b9e-a314-bf496c853710-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.945160] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.945160] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.059734] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.060142] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.060489] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1009.323611] env[61906]: DEBUG nova.compute.manager [req-11137475-c8d0-4129-bf51-1f42d9868b29 req-1df91595-2494-4175-95ba-7965330e894b service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Received event network-changed-fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.323611] env[61906]: DEBUG nova.compute.manager [req-11137475-c8d0-4129-bf51-1f42d9868b29 req-1df91595-2494-4175-95ba-7965330e894b service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Refreshing instance network info cache due to event network-changed-fc9835ce-e30c-4b58-8804-1062f5d166de. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1009.324333] env[61906]: DEBUG oslo_concurrency.lockutils [req-11137475-c8d0-4129-bf51-1f42d9868b29 req-1df91595-2494-4175-95ba-7965330e894b service nova] Acquiring lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.324333] env[61906]: DEBUG oslo_concurrency.lockutils [req-11137475-c8d0-4129-bf51-1f42d9868b29 req-1df91595-2494-4175-95ba-7965330e894b service nova] Acquired lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.327905] env[61906]: DEBUG nova.network.neutron [req-11137475-c8d0-4129-bf51-1f42d9868b29 req-1df91595-2494-4175-95ba-7965330e894b service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Refreshing network info cache for port fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1009.579014] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "91dc442c-0272-419d-8ef9-06e279a3ea61" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.579336] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "91dc442c-0272-419d-8ef9-06e279a3ea61" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.002794] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.002997] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.003211] env[61906]: DEBUG nova.network.neutron [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1010.061544] env[61906]: DEBUG nova.network.neutron [req-11137475-c8d0-4129-bf51-1f42d9868b29 req-1df91595-2494-4175-95ba-7965330e894b service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updated VIF entry in instance network info cache for port fc9835ce-e30c-4b58-8804-1062f5d166de. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1010.062955] env[61906]: DEBUG nova.network.neutron [req-11137475-c8d0-4129-bf51-1f42d9868b29 req-1df91595-2494-4175-95ba-7965330e894b service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updating instance_info_cache with network_info: [{"id": "fc9835ce-e30c-4b58-8804-1062f5d166de", "address": "fa:16:3e:b8:59:0b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc9835ce-e3", "ovs_interfaceid": "fc9835ce-e30c-4b58-8804-1062f5d166de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.073381] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.083108] env[61906]: DEBUG nova.compute.manager [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1010.499755] env[61906]: DEBUG nova.compute.manager [req-cf7ac083-a0d6-4baf-916c-8644ac56941a req-a8d4dc0e-f29f-4378-b355-7459dc30d8b3 service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Received event network-changed-a6bc2330-c2b4-4d0b-8178-f827316508a3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.499956] env[61906]: DEBUG nova.compute.manager [req-cf7ac083-a0d6-4baf-916c-8644ac56941a req-a8d4dc0e-f29f-4378-b355-7459dc30d8b3 service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Refreshing instance network info cache due to event network-changed-a6bc2330-c2b4-4d0b-8178-f827316508a3. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1010.500202] env[61906]: DEBUG oslo_concurrency.lockutils [req-cf7ac083-a0d6-4baf-916c-8644ac56941a req-a8d4dc0e-f29f-4378-b355-7459dc30d8b3 service nova] Acquiring lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.500399] env[61906]: DEBUG oslo_concurrency.lockutils [req-cf7ac083-a0d6-4baf-916c-8644ac56941a req-a8d4dc0e-f29f-4378-b355-7459dc30d8b3 service nova] Acquired lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.500524] env[61906]: DEBUG nova.network.neutron [req-cf7ac083-a0d6-4baf-916c-8644ac56941a req-a8d4dc0e-f29f-4378-b355-7459dc30d8b3 service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Refreshing network info cache for port a6bc2330-c2b4-4d0b-8178-f827316508a3 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1010.565046] env[61906]: DEBUG oslo_concurrency.lockutils [req-11137475-c8d0-4129-bf51-1f42d9868b29 req-1df91595-2494-4175-95ba-7965330e894b service nova] Releasing lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.609242] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.609760] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.613131] env[61906]: INFO nova.compute.claims [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.748698] env[61906]: DEBUG nova.network.neutron [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance_info_cache with network_info: [{"id": "d803c8e8-7145-412c-aa3e-2545666742eb", "address": "fa:16:3e:46:59:77", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd803c8e8-71", "ovs_interfaceid": "d803c8e8-7145-412c-aa3e-2545666742eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.204535] env[61906]: DEBUG nova.network.neutron [req-cf7ac083-a0d6-4baf-916c-8644ac56941a req-a8d4dc0e-f29f-4378-b355-7459dc30d8b3 service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updated VIF entry in instance network info cache for port a6bc2330-c2b4-4d0b-8178-f827316508a3. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1011.204898] env[61906]: DEBUG nova.network.neutron [req-cf7ac083-a0d6-4baf-916c-8644ac56941a req-a8d4dc0e-f29f-4378-b355-7459dc30d8b3 service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updating instance_info_cache with network_info: [{"id": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "address": "fa:16:3e:87:df:8b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6bc2330-c2", "ovs_interfaceid": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.253442] env[61906]: DEBUG oslo_concurrency.lockutils [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.257052] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquired lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.257220] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: e133605d-c630-4b9e-a314-bf496c853710] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1011.349815] env[61906]: DEBUG nova.compute.manager [req-6ad9f76e-fe91-4a78-a22e-d7c73eac8bdf req-b92fcf0a-40f0-472d-9464-f653f2546420 service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Received event network-changed-a6bc2330-c2b4-4d0b-8178-f827316508a3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.349999] env[61906]: DEBUG nova.compute.manager [req-6ad9f76e-fe91-4a78-a22e-d7c73eac8bdf req-b92fcf0a-40f0-472d-9464-f653f2546420 service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Refreshing instance network info cache due to event network-changed-a6bc2330-c2b4-4d0b-8178-f827316508a3. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1011.350209] env[61906]: DEBUG oslo_concurrency.lockutils [req-6ad9f76e-fe91-4a78-a22e-d7c73eac8bdf req-b92fcf0a-40f0-472d-9464-f653f2546420 service nova] Acquiring lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.705519] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96bd7e37-432a-4de9-9dc1-83974179c639 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.708523] env[61906]: DEBUG oslo_concurrency.lockutils [req-cf7ac083-a0d6-4baf-916c-8644ac56941a req-a8d4dc0e-f29f-4378-b355-7459dc30d8b3 service nova] Releasing lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.708921] env[61906]: DEBUG oslo_concurrency.lockutils [req-6ad9f76e-fe91-4a78-a22e-d7c73eac8bdf req-b92fcf0a-40f0-472d-9464-f653f2546420 service nova] Acquired lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.709147] env[61906]: DEBUG nova.network.neutron [req-6ad9f76e-fe91-4a78-a22e-d7c73eac8bdf req-b92fcf0a-40f0-472d-9464-f653f2546420 service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Refreshing network info cache for port a6bc2330-c2b4-4d0b-8178-f827316508a3 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1011.715580] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb048ee-ad0d-4e2f-b7eb-babca9dab889 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.746331] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13358f9-8cbb-41b7-955e-964fa096e5a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.753613] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e44e68-3262-4207-86f6-e1ece4c8e54f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.772703] env[61906]: DEBUG nova.compute.provider_tree [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1011.784011] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef548450-2c4f-4563-b92d-485a9193b41c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.803314] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e5ea5a-df99-459e-bfcc-4540532cd90a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.810273] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance 'e133605d-c630-4b9e-a314-bf496c853710' progress to 83 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1012.291603] env[61906]: ERROR nova.scheduler.client.report [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [req-7b89fccf-4fd8-4ed1-a76b-12f6590a0fb6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6e524f5d-985e-41eb-a7ae-996be3ae1c1a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7b89fccf-4fd8-4ed1-a76b-12f6590a0fb6"}]} [ 1012.309516] env[61906]: DEBUG nova.scheduler.client.report [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Refreshing inventories for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1012.317811] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1012.318173] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c24ae7d-043f-4631-911b-d4b205825c52 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.323105] env[61906]: DEBUG nova.scheduler.client.report [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updating ProviderTree inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1012.323332] env[61906]: DEBUG nova.compute.provider_tree [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1012.326368] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1012.326368] env[61906]: value = "task-1357046" [ 1012.326368] env[61906]: _type = "Task" [ 1012.326368] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.334053] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357046, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.336055] env[61906]: DEBUG nova.scheduler.client.report [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Refreshing aggregate associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1012.355663] env[61906]: DEBUG nova.scheduler.client.report [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Refreshing trait associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1012.416748] env[61906]: DEBUG nova.network.neutron [req-6ad9f76e-fe91-4a78-a22e-d7c73eac8bdf req-b92fcf0a-40f0-472d-9464-f653f2546420 service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updated VIF entry in instance network info cache for port a6bc2330-c2b4-4d0b-8178-f827316508a3. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1012.417138] env[61906]: DEBUG nova.network.neutron [req-6ad9f76e-fe91-4a78-a22e-d7c73eac8bdf req-b92fcf0a-40f0-472d-9464-f653f2546420 service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updating instance_info_cache with network_info: [{"id": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "address": "fa:16:3e:87:df:8b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6bc2330-c2", "ovs_interfaceid": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.434929] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e622544f-8d0b-456c-b6c1-d9c1dca4291a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.445132] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2b0acc-4c96-42cd-bf9c-8af635214911 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.478915] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89790f8-1319-43c4-af3b-69d2857b348c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.486614] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557b5071-e010-4bc9-8fe3-805ed9d4c56d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.499712] env[61906]: DEBUG nova.compute.provider_tree [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1012.523497] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance_info_cache with network_info: [{"id": "d803c8e8-7145-412c-aa3e-2545666742eb", "address": "fa:16:3e:46:59:77", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd803c8e8-71", "ovs_interfaceid": "d803c8e8-7145-412c-aa3e-2545666742eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.528265] env[61906]: DEBUG nova.compute.manager [req-7d98e17f-12e4-41db-9148-477e356bc03a req-4cb56038-05e4-4131-acb9-deba749fba87 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Received event network-changed-fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.528913] env[61906]: DEBUG nova.compute.manager [req-7d98e17f-12e4-41db-9148-477e356bc03a req-4cb56038-05e4-4131-acb9-deba749fba87 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Refreshing instance network info cache due to event network-changed-fc9835ce-e30c-4b58-8804-1062f5d166de. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1012.528913] env[61906]: DEBUG oslo_concurrency.lockutils [req-7d98e17f-12e4-41db-9148-477e356bc03a req-4cb56038-05e4-4131-acb9-deba749fba87 service nova] Acquiring lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.528913] env[61906]: DEBUG oslo_concurrency.lockutils [req-7d98e17f-12e4-41db-9148-477e356bc03a req-4cb56038-05e4-4131-acb9-deba749fba87 service nova] Acquired lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.529023] env[61906]: DEBUG nova.network.neutron [req-7d98e17f-12e4-41db-9148-477e356bc03a req-4cb56038-05e4-4131-acb9-deba749fba87 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Refreshing network info cache for port fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1012.836114] env[61906]: DEBUG oslo_vmware.api [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357046, 'name': PowerOnVM_Task, 'duration_secs': 0.437213} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.836438] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1012.836629] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-da01aaf4-77f5-47aa-948e-bc9674f66589 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance 'e133605d-c630-4b9e-a314-bf496c853710' progress to 100 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1012.919682] env[61906]: DEBUG oslo_concurrency.lockutils [req-6ad9f76e-fe91-4a78-a22e-d7c73eac8bdf req-b92fcf0a-40f0-472d-9464-f653f2546420 service nova] Releasing lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.025889] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Releasing lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.026259] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1013.026511] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1013.026784] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1013.027345] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1013.027345] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1013.027345] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1013.027482] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1013.028276] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1013.028276] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1013.034162] env[61906]: DEBUG nova.scheduler.client.report [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updated inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with generation 119 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1013.034427] env[61906]: DEBUG nova.compute.provider_tree [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updating resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a generation from 119 to 120 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1013.034614] env[61906]: DEBUG nova.compute.provider_tree [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1013.241611] env[61906]: DEBUG nova.network.neutron [req-7d98e17f-12e4-41db-9148-477e356bc03a req-4cb56038-05e4-4131-acb9-deba749fba87 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updated VIF entry in instance network info cache for port fc9835ce-e30c-4b58-8804-1062f5d166de. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1013.242015] env[61906]: DEBUG nova.network.neutron [req-7d98e17f-12e4-41db-9148-477e356bc03a req-4cb56038-05e4-4131-acb9-deba749fba87 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updating instance_info_cache with network_info: [{"id": "fc9835ce-e30c-4b58-8804-1062f5d166de", "address": "fa:16:3e:b8:59:0b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc9835ce-e3", "ovs_interfaceid": "fc9835ce-e30c-4b58-8804-1062f5d166de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.530103] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.540193] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.931s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.540815] env[61906]: DEBUG nova.compute.manager [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1013.543979] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.014s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.544162] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.544338] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1013.545271] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8754867e-9eb4-48bf-ba1d-372db9cec6be {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.553677] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b08c45-61ab-43d8-8265-c39edbf51755 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.568792] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31118285-15df-4803-92ef-97c6817adc90 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.575240] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b258adf1-d335-45f5-8d68-eb7ec8ae9fa1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.603181] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180278MB free_disk=150GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1013.603303] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.603489] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.744620] env[61906]: DEBUG oslo_concurrency.lockutils [req-7d98e17f-12e4-41db-9148-477e356bc03a req-4cb56038-05e4-4131-acb9-deba749fba87 service nova] Releasing lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.045957] env[61906]: DEBUG nova.compute.utils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1014.047890] env[61906]: DEBUG nova.compute.manager [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1014.047890] env[61906]: DEBUG nova.network.neutron [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1014.091694] env[61906]: DEBUG nova.policy [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e28284ad5ce450597ef12c65c902e1c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ac4f2c9065c46068c7837eb85cd7e20', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1014.402689] env[61906]: DEBUG nova.network.neutron [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Successfully created port: a10b72b0-6aab-4516-9e87-ce2ea1c1ae15 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1014.550415] env[61906]: DEBUG nova.compute.manager [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1014.611303] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Applying migration context for instance e133605d-c630-4b9e-a314-bf496c853710 as it has an incoming, in-progress migration f9f0d985-c6eb-4adc-adab-c9f39bfc4d5c. Migration status is finished {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1014.612198] env[61906]: INFO nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating resource usage from migration f9f0d985-c6eb-4adc-adab-c9f39bfc4d5c [ 1014.634368] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 3ea24d58-ffbd-41d4-a010-978a395ad2fc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1014.634518] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance d8c1fc04-d5ac-4c78-95e1-e135b22819c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1014.634647] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 729f5565-8cda-4bee-8644-fb94a06034fb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1014.634770] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Migration f9f0d985-c6eb-4adc-adab-c9f39bfc4d5c is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1014.634889] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance e133605d-c630-4b9e-a314-bf496c853710 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1014.635104] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 91dc442c-0272-419d-8ef9-06e279a3ea61 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1014.635213] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1014.635352] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1728MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1014.711096] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a110d03-dfd6-4b99-ade8-ab5b97b7a9e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.719220] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ecd0a6e-4b8f-4ec4-af7a-d0cc1bdf25e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.749862] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db3e7be0-3638-487d-8f90-d25359ae0f6c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.757429] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b325f8-95f4-46a6-99b1-4e7cb6df1212 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.770970] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "e133605d-c630-4b9e-a314-bf496c853710" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.771246] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.771460] env[61906]: DEBUG nova.compute.manager [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Going to confirm migration 3 {{(pid=61906) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1014.772935] env[61906]: DEBUG nova.compute.provider_tree [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.278252] env[61906]: DEBUG nova.scheduler.client.report [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.351885] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.352084] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.352803] env[61906]: DEBUG nova.network.neutron [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1015.352803] env[61906]: DEBUG nova.objects.instance [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lazy-loading 'info_cache' on Instance uuid e133605d-c630-4b9e-a314-bf496c853710 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.559355] env[61906]: DEBUG nova.compute.manager [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1015.584144] env[61906]: DEBUG nova.virt.hardware [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1015.584410] env[61906]: DEBUG nova.virt.hardware [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1015.584581] env[61906]: DEBUG nova.virt.hardware [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.584781] env[61906]: DEBUG nova.virt.hardware [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1015.584934] env[61906]: DEBUG nova.virt.hardware [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.585101] env[61906]: DEBUG nova.virt.hardware [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1015.585317] env[61906]: DEBUG nova.virt.hardware [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1015.585482] env[61906]: DEBUG nova.virt.hardware [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1015.585654] env[61906]: DEBUG nova.virt.hardware [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1015.585824] env[61906]: DEBUG nova.virt.hardware [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1015.585999] env[61906]: DEBUG nova.virt.hardware [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1015.586871] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fae0e5a-a908-4dc4-bd42-99aa7b293800 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.595073] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a81c8ecf-b9d8-4739-a0ee-82b9e4133d69 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.782778] env[61906]: DEBUG nova.compute.manager [req-e11be9d5-9c45-497d-af5b-81b9087acbe0 req-e8139846-0c21-477d-ba87-04586f23a7e9 service nova] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Received event network-vif-plugged-a10b72b0-6aab-4516-9e87-ce2ea1c1ae15 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.783059] env[61906]: DEBUG oslo_concurrency.lockutils [req-e11be9d5-9c45-497d-af5b-81b9087acbe0 req-e8139846-0c21-477d-ba87-04586f23a7e9 service nova] Acquiring lock "91dc442c-0272-419d-8ef9-06e279a3ea61-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.783284] env[61906]: DEBUG oslo_concurrency.lockutils [req-e11be9d5-9c45-497d-af5b-81b9087acbe0 req-e8139846-0c21-477d-ba87-04586f23a7e9 service nova] Lock "91dc442c-0272-419d-8ef9-06e279a3ea61-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.783459] env[61906]: DEBUG oslo_concurrency.lockutils [req-e11be9d5-9c45-497d-af5b-81b9087acbe0 req-e8139846-0c21-477d-ba87-04586f23a7e9 service nova] Lock "91dc442c-0272-419d-8ef9-06e279a3ea61-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.783630] env[61906]: DEBUG nova.compute.manager [req-e11be9d5-9c45-497d-af5b-81b9087acbe0 req-e8139846-0c21-477d-ba87-04586f23a7e9 service nova] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] No waiting events found dispatching network-vif-plugged-a10b72b0-6aab-4516-9e87-ce2ea1c1ae15 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1015.783801] env[61906]: WARNING nova.compute.manager [req-e11be9d5-9c45-497d-af5b-81b9087acbe0 req-e8139846-0c21-477d-ba87-04586f23a7e9 service nova] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Received unexpected event network-vif-plugged-a10b72b0-6aab-4516-9e87-ce2ea1c1ae15 for instance with vm_state building and task_state spawning. [ 1015.784881] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1015.785065] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.182s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.864359] env[61906]: DEBUG nova.network.neutron [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Successfully updated port: a10b72b0-6aab-4516-9e87-ce2ea1c1ae15 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.367797] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "refresh_cache-91dc442c-0272-419d-8ef9-06e279a3ea61" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.368194] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "refresh_cache-91dc442c-0272-419d-8ef9-06e279a3ea61" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.368194] env[61906]: DEBUG nova.network.neutron [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1016.603287] env[61906]: DEBUG nova.network.neutron [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance_info_cache with network_info: [{"id": "d803c8e8-7145-412c-aa3e-2545666742eb", "address": "fa:16:3e:46:59:77", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd803c8e8-71", "ovs_interfaceid": "d803c8e8-7145-412c-aa3e-2545666742eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.901664] env[61906]: DEBUG nova.network.neutron [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1017.105992] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-e133605d-c630-4b9e-a314-bf496c853710" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.106311] env[61906]: DEBUG nova.objects.instance [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lazy-loading 'migration_context' on Instance uuid e133605d-c630-4b9e-a314-bf496c853710 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.185310] env[61906]: DEBUG nova.network.neutron [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Updating instance_info_cache with network_info: [{"id": "a10b72b0-6aab-4516-9e87-ce2ea1c1ae15", "address": "fa:16:3e:6c:6d:7a", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa10b72b0-6a", "ovs_interfaceid": "a10b72b0-6aab-4516-9e87-ce2ea1c1ae15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.608964] env[61906]: DEBUG nova.objects.base [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1017.610011] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a38db73-3e8f-4c2a-9600-d816cb3d0397 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.629278] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6db0385c-2d93-4a28-af51-ecf5144537cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.634950] env[61906]: DEBUG oslo_vmware.api [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1017.634950] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ae7d96-6548-aaff-1337-168e5cc2a9c7" [ 1017.634950] env[61906]: _type = "Task" [ 1017.634950] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.643957] env[61906]: DEBUG oslo_vmware.api [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ae7d96-6548-aaff-1337-168e5cc2a9c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.687745] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "refresh_cache-91dc442c-0272-419d-8ef9-06e279a3ea61" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.688122] env[61906]: DEBUG nova.compute.manager [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Instance network_info: |[{"id": "a10b72b0-6aab-4516-9e87-ce2ea1c1ae15", "address": "fa:16:3e:6c:6d:7a", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa10b72b0-6a", "ovs_interfaceid": "a10b72b0-6aab-4516-9e87-ce2ea1c1ae15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1017.688571] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:6d:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1895250-76cc-41f7-b7f8-2e5679494607', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a10b72b0-6aab-4516-9e87-ce2ea1c1ae15', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.697763] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Creating folder: Project (3ac4f2c9065c46068c7837eb85cd7e20). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1017.697984] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d18c6713-3a79-4555-beae-8c25b1398d58 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.708686] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Created folder: Project (3ac4f2c9065c46068c7837eb85cd7e20) in parent group-v288914. [ 1017.708884] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Creating folder: Instances. Parent ref: group-v289037. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1017.709129] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3e96dd15-0f73-4b77-b999-8fe072730ca7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.717881] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Created folder: Instances in parent group-v289037. [ 1017.718136] env[61906]: DEBUG oslo.service.loopingcall [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.718330] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1017.718530] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9f97ba44-975b-4eed-b5e6-39b37d71d465 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.736408] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.736408] env[61906]: value = "task-1357049" [ 1017.736408] env[61906]: _type = "Task" [ 1017.736408] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.743448] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357049, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.810781] env[61906]: DEBUG nova.compute.manager [req-2f855bee-2069-47a2-b6bc-29dd4970d979 req-2d7596fd-2583-49bb-9ff9-2e7427043421 service nova] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Received event network-changed-a10b72b0-6aab-4516-9e87-ce2ea1c1ae15 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.811070] env[61906]: DEBUG nova.compute.manager [req-2f855bee-2069-47a2-b6bc-29dd4970d979 req-2d7596fd-2583-49bb-9ff9-2e7427043421 service nova] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Refreshing instance network info cache due to event network-changed-a10b72b0-6aab-4516-9e87-ce2ea1c1ae15. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1017.811337] env[61906]: DEBUG oslo_concurrency.lockutils [req-2f855bee-2069-47a2-b6bc-29dd4970d979 req-2d7596fd-2583-49bb-9ff9-2e7427043421 service nova] Acquiring lock "refresh_cache-91dc442c-0272-419d-8ef9-06e279a3ea61" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.811907] env[61906]: DEBUG oslo_concurrency.lockutils [req-2f855bee-2069-47a2-b6bc-29dd4970d979 req-2d7596fd-2583-49bb-9ff9-2e7427043421 service nova] Acquired lock "refresh_cache-91dc442c-0272-419d-8ef9-06e279a3ea61" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.811907] env[61906]: DEBUG nova.network.neutron [req-2f855bee-2069-47a2-b6bc-29dd4970d979 req-2d7596fd-2583-49bb-9ff9-2e7427043421 service nova] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Refreshing network info cache for port a10b72b0-6aab-4516-9e87-ce2ea1c1ae15 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1018.145111] env[61906]: DEBUG oslo_vmware.api [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ae7d96-6548-aaff-1337-168e5cc2a9c7, 'name': SearchDatastore_Task, 'duration_secs': 0.008724} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.145441] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.145680] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.245646] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357049, 'name': CreateVM_Task, 'duration_secs': 0.304966} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.245802] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1018.246513] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.246685] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.247024] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.247312] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f67f92c-e2a5-4c8f-8451-9592c304eb80 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.251478] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1018.251478] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52356389-28b0-f001-2109-ec7a6d2e4273" [ 1018.251478] env[61906]: _type = "Task" [ 1018.251478] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.258435] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52356389-28b0-f001-2109-ec7a6d2e4273, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.495514] env[61906]: DEBUG nova.network.neutron [req-2f855bee-2069-47a2-b6bc-29dd4970d979 req-2d7596fd-2583-49bb-9ff9-2e7427043421 service nova] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Updated VIF entry in instance network info cache for port a10b72b0-6aab-4516-9e87-ce2ea1c1ae15. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1018.495887] env[61906]: DEBUG nova.network.neutron [req-2f855bee-2069-47a2-b6bc-29dd4970d979 req-2d7596fd-2583-49bb-9ff9-2e7427043421 service nova] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Updating instance_info_cache with network_info: [{"id": "a10b72b0-6aab-4516-9e87-ce2ea1c1ae15", "address": "fa:16:3e:6c:6d:7a", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa10b72b0-6a", "ovs_interfaceid": "a10b72b0-6aab-4516-9e87-ce2ea1c1ae15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.735577] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f56acf8-de1f-4018-830c-a131a3f0c9b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.743009] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b279f42d-82b9-4130-bbbb-c3f122d5493c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.776940] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc965ae-1ef2-447a-8b91-16cc55c5248e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.784764] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52356389-28b0-f001-2109-ec7a6d2e4273, 'name': SearchDatastore_Task, 'duration_secs': 0.008804} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.786747] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.786994] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.787254] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.787411] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.787625] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.787948] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-319408e0-291e-41d5-94b3-5fb9b7ea2d04 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.790531] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f938df3-5e7c-4b0f-80ec-549da6a471e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.803816] env[61906]: DEBUG nova.compute.provider_tree [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.806085] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.806270] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1018.807178] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34174059-7ae1-4602-998c-465cc1deb9b7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.812303] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1018.812303] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521dee7f-1736-7f68-769c-80084af07de9" [ 1018.812303] env[61906]: _type = "Task" [ 1018.812303] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.820616] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521dee7f-1736-7f68-769c-80084af07de9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.999207] env[61906]: DEBUG oslo_concurrency.lockutils [req-2f855bee-2069-47a2-b6bc-29dd4970d979 req-2d7596fd-2583-49bb-9ff9-2e7427043421 service nova] Releasing lock "refresh_cache-91dc442c-0272-419d-8ef9-06e279a3ea61" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.308888] env[61906]: DEBUG nova.scheduler.client.report [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.322963] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521dee7f-1736-7f68-769c-80084af07de9, 'name': SearchDatastore_Task, 'duration_secs': 0.008737} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.324295] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-328d5ce3-c248-426e-b5c9-b38b151af938 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.329677] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1019.329677] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b600e8-80dd-5f9d-591f-72ac54ef93ac" [ 1019.329677] env[61906]: _type = "Task" [ 1019.329677] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.337064] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b600e8-80dd-5f9d-591f-72ac54ef93ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.840615] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b600e8-80dd-5f9d-591f-72ac54ef93ac, 'name': SearchDatastore_Task, 'duration_secs': 0.008939} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.840924] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.841209] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 91dc442c-0272-419d-8ef9-06e279a3ea61/91dc442c-0272-419d-8ef9-06e279a3ea61.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1019.841481] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a695f2a7-df67-4a65-8c9b-bea0819a57e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.848209] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1019.848209] env[61906]: value = "task-1357050" [ 1019.848209] env[61906]: _type = "Task" [ 1019.848209] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.856292] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357050, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.318858] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.173s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.359371] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357050, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446575} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.359649] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 91dc442c-0272-419d-8ef9-06e279a3ea61/91dc442c-0272-419d-8ef9-06e279a3ea61.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1020.359879] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.360401] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bdba5278-044f-48b9-bea0-e21b8f728542 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.368175] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1020.368175] env[61906]: value = "task-1357051" [ 1020.368175] env[61906]: _type = "Task" [ 1020.368175] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.376628] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357051, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.875071] env[61906]: INFO nova.scheduler.client.report [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleted allocation for migration f9f0d985-c6eb-4adc-adab-c9f39bfc4d5c [ 1020.881592] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357051, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057412} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.882105] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.882866] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155d1291-e954-4c75-b19f-1b48855b1f60 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.904017] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 91dc442c-0272-419d-8ef9-06e279a3ea61/91dc442c-0272-419d-8ef9-06e279a3ea61.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.904327] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d92c552a-1e17-4e52-a710-5e18f124f44f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.924836] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1020.924836] env[61906]: value = "task-1357052" [ 1020.924836] env[61906]: _type = "Task" [ 1020.924836] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.933809] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357052, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.384608] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d5c6e472-766e-4b45-9553-2fc9643d1a0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.613s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.435298] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357052, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.815757] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "e133605d-c630-4b9e-a314-bf496c853710" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.816079] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.816312] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "e133605d-c630-4b9e-a314-bf496c853710-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.816510] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.816692] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.819109] env[61906]: INFO nova.compute.manager [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Terminating instance [ 1021.820938] env[61906]: DEBUG nova.compute.manager [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1021.821148] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1021.821978] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317463d6-7a92-45e6-a265-2c2060ff9a50 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.829733] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.829961] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3e235373-2943-41e5-987e-d6f5a02c9fa9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.836375] env[61906]: DEBUG oslo_vmware.api [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1021.836375] env[61906]: value = "task-1357053" [ 1021.836375] env[61906]: _type = "Task" [ 1021.836375] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.843920] env[61906]: DEBUG oslo_vmware.api [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357053, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.935904] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357052, 'name': ReconfigVM_Task, 'duration_secs': 0.750515} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.936295] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 91dc442c-0272-419d-8ef9-06e279a3ea61/91dc442c-0272-419d-8ef9-06e279a3ea61.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.936819] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e43e8b26-0ff4-4841-bacb-086dd5d2ce91 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.943145] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1021.943145] env[61906]: value = "task-1357054" [ 1021.943145] env[61906]: _type = "Task" [ 1021.943145] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.951480] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357054, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.345795] env[61906]: DEBUG oslo_vmware.api [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357053, 'name': PowerOffVM_Task, 'duration_secs': 0.176653} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.346036] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1022.346220] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1022.346468] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-75c1797f-34b4-4deb-8310-96dff9f61744 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.425379] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1022.425607] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1022.425799] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleting the datastore file [datastore1] e133605d-c630-4b9e-a314-bf496c853710 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.426088] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-005c25f5-e2c6-4f70-8a16-91ee12711bf4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.431784] env[61906]: DEBUG oslo_vmware.api [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1022.431784] env[61906]: value = "task-1357056" [ 1022.431784] env[61906]: _type = "Task" [ 1022.431784] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.440979] env[61906]: DEBUG oslo_vmware.api [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.451414] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357054, 'name': Rename_Task, 'duration_secs': 0.135962} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.451673] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1022.451929] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea32d1b3-f100-427b-91bd-67463265b0f4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.457600] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1022.457600] env[61906]: value = "task-1357057" [ 1022.457600] env[61906]: _type = "Task" [ 1022.457600] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.464574] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357057, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.738662] env[61906]: DEBUG oslo_concurrency.lockutils [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "interface-3ea24d58-ffbd-41d4-a010-978a395ad2fc-7ce89428-70fc-4bff-a80f-235a2dc0a2a0" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.738956] env[61906]: DEBUG oslo_concurrency.lockutils [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-3ea24d58-ffbd-41d4-a010-978a395ad2fc-7ce89428-70fc-4bff-a80f-235a2dc0a2a0" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.739373] env[61906]: DEBUG nova.objects.instance [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'flavor' on Instance uuid 3ea24d58-ffbd-41d4-a010-978a395ad2fc {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.944089] env[61906]: DEBUG oslo_vmware.api [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357056, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137889} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.944488] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1022.944739] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1022.945217] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1022.945343] env[61906]: INFO nova.compute.manager [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: e133605d-c630-4b9e-a314-bf496c853710] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1022.945626] env[61906]: DEBUG oslo.service.loopingcall [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.945889] env[61906]: DEBUG nova.compute.manager [-] [instance: e133605d-c630-4b9e-a314-bf496c853710] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1022.946039] env[61906]: DEBUG nova.network.neutron [-] [instance: e133605d-c630-4b9e-a314-bf496c853710] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1022.967187] env[61906]: DEBUG oslo_vmware.api [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357057, 'name': PowerOnVM_Task, 'duration_secs': 0.453116} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.967446] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1022.967647] env[61906]: INFO nova.compute.manager [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Took 7.41 seconds to spawn the instance on the hypervisor. [ 1022.967833] env[61906]: DEBUG nova.compute.manager [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.968624] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e40908-2e27-4274-bfc4-5e5210dc18d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.359123] env[61906]: DEBUG nova.objects.instance [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'pci_requests' on Instance uuid 3ea24d58-ffbd-41d4-a010-978a395ad2fc {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.486064] env[61906]: INFO nova.compute.manager [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Took 12.90 seconds to build instance. [ 1023.695468] env[61906]: DEBUG nova.compute.manager [req-7d62e36f-ef4e-45bd-8a47-3beddc4da0f8 req-c6d388c7-1047-4db3-baee-ebd35c6e3940 service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Received event network-vif-deleted-d803c8e8-7145-412c-aa3e-2545666742eb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1023.695468] env[61906]: INFO nova.compute.manager [req-7d62e36f-ef4e-45bd-8a47-3beddc4da0f8 req-c6d388c7-1047-4db3-baee-ebd35c6e3940 service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Neutron deleted interface d803c8e8-7145-412c-aa3e-2545666742eb; detaching it from the instance and deleting it from the info cache [ 1023.695468] env[61906]: DEBUG nova.network.neutron [req-7d62e36f-ef4e-45bd-8a47-3beddc4da0f8 req-c6d388c7-1047-4db3-baee-ebd35c6e3940 service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.862073] env[61906]: DEBUG nova.objects.base [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Object Instance<3ea24d58-ffbd-41d4-a010-978a395ad2fc> lazy-loaded attributes: flavor,pci_requests {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1023.862439] env[61906]: DEBUG nova.network.neutron [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1023.988579] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2c166de2-46dd-4afd-b68b-85a6b37d8194 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "91dc442c-0272-419d-8ef9-06e279a3ea61" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.409s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.031070] env[61906]: DEBUG nova.policy [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73219f5e96f64d1da86a19da9b18b06a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5533be74874b4094b7ef0f98a121f60b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1024.039800] env[61906]: DEBUG nova.network.neutron [-] [instance: e133605d-c630-4b9e-a314-bf496c853710] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.201360] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c46ac4c-8451-4926-9db1-a4f8d66e9a3c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.213657] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d15464d-377b-4c67-a055-7ccc736c41e2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.249910] env[61906]: DEBUG nova.compute.manager [req-7d62e36f-ef4e-45bd-8a47-3beddc4da0f8 req-c6d388c7-1047-4db3-baee-ebd35c6e3940 service nova] [instance: e133605d-c630-4b9e-a314-bf496c853710] Detach interface failed, port_id=d803c8e8-7145-412c-aa3e-2545666742eb, reason: Instance e133605d-c630-4b9e-a314-bf496c853710 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1024.541668] env[61906]: INFO nova.compute.manager [-] [instance: e133605d-c630-4b9e-a314-bf496c853710] Took 1.60 seconds to deallocate network for instance. [ 1024.795329] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "1e1a742f-824f-4118-82fb-cafce626e0ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.795548] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "1e1a742f-824f-4118-82fb-cafce626e0ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.048745] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.049088] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.049204] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.066989] env[61906]: INFO nova.scheduler.client.report [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleted allocations for instance e133605d-c630-4b9e-a314-bf496c853710 [ 1025.297549] env[61906]: DEBUG nova.compute.manager [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1025.575992] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0a0a4858-29ad-4186-8fcd-3f0a6a9f6e0f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "e133605d-c630-4b9e-a314-bf496c853710" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.760s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.629492] env[61906]: DEBUG nova.network.neutron [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Successfully updated port: 7ce89428-70fc-4bff-a80f-235a2dc0a2a0 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1025.720896] env[61906]: DEBUG nova.compute.manager [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Received event network-vif-plugged-7ce89428-70fc-4bff-a80f-235a2dc0a2a0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1025.721033] env[61906]: DEBUG oslo_concurrency.lockutils [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] Acquiring lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.721274] env[61906]: DEBUG oslo_concurrency.lockutils [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] Lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.721471] env[61906]: DEBUG oslo_concurrency.lockutils [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] Lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.721647] env[61906]: DEBUG nova.compute.manager [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] No waiting events found dispatching network-vif-plugged-7ce89428-70fc-4bff-a80f-235a2dc0a2a0 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1025.721821] env[61906]: WARNING nova.compute.manager [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Received unexpected event network-vif-plugged-7ce89428-70fc-4bff-a80f-235a2dc0a2a0 for instance with vm_state active and task_state None. [ 1025.721988] env[61906]: DEBUG nova.compute.manager [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Received event network-changed-7ce89428-70fc-4bff-a80f-235a2dc0a2a0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1025.722174] env[61906]: DEBUG nova.compute.manager [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Refreshing instance network info cache due to event network-changed-7ce89428-70fc-4bff-a80f-235a2dc0a2a0. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1025.722359] env[61906]: DEBUG oslo_concurrency.lockutils [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] Acquiring lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.722498] env[61906]: DEBUG oslo_concurrency.lockutils [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] Acquired lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.722655] env[61906]: DEBUG nova.network.neutron [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Refreshing network info cache for port 7ce89428-70fc-4bff-a80f-235a2dc0a2a0 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1025.818505] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.818840] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.820844] env[61906]: INFO nova.compute.claims [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.132995] env[61906]: DEBUG oslo_concurrency.lockutils [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.456239] env[61906]: DEBUG nova.network.neutron [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Added VIF to instance network info cache for port 7ce89428-70fc-4bff-a80f-235a2dc0a2a0. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3546}} [ 1026.456670] env[61906]: DEBUG nova.network.neutron [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updating instance_info_cache with network_info: [{"id": "fc9835ce-e30c-4b58-8804-1062f5d166de", "address": "fa:16:3e:b8:59:0b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc9835ce-e3", "ovs_interfaceid": "fc9835ce-e30c-4b58-8804-1062f5d166de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7ce89428-70fc-4bff-a80f-235a2dc0a2a0", "address": "fa:16:3e:1f:65:b9", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ce89428-70", "ovs_interfaceid": "7ce89428-70fc-4bff-a80f-235a2dc0a2a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.767213] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.767458] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.901651] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb81c4ef-8073-447b-9eab-1ad045c9e0af {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.909503] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d38c1574-2cd3-402f-8b07-3096aef8a841 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.939264] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd1abcee-aee8-4201-a792-c8859a7000a3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.947280] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bfb5ef9-7b09-440e-b100-675f1d191779 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.960578] env[61906]: DEBUG oslo_concurrency.lockutils [req-cfa65ff1-d466-4667-a5cf-53ff776b6ea8 req-acf562ce-6c2f-4c34-83f0-500192afb4cb service nova] Releasing lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.961296] env[61906]: DEBUG nova.compute.provider_tree [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.962632] env[61906]: DEBUG oslo_concurrency.lockutils [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.962822] env[61906]: DEBUG nova.network.neutron [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.269847] env[61906]: DEBUG nova.compute.manager [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1027.466068] env[61906]: DEBUG nova.scheduler.client.report [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.498990] env[61906]: WARNING nova.network.neutron [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] b12c63e1-bd61-4da3-8d37-4a89206ba0f2 already exists in list: networks containing: ['b12c63e1-bd61-4da3-8d37-4a89206ba0f2']. ignoring it [ 1027.499232] env[61906]: WARNING nova.network.neutron [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] b12c63e1-bd61-4da3-8d37-4a89206ba0f2 already exists in list: networks containing: ['b12c63e1-bd61-4da3-8d37-4a89206ba0f2']. ignoring it [ 1027.499408] env[61906]: WARNING nova.network.neutron [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] 7ce89428-70fc-4bff-a80f-235a2dc0a2a0 already exists in list: port_ids containing: ['7ce89428-70fc-4bff-a80f-235a2dc0a2a0']. ignoring it [ 1027.768289] env[61906]: DEBUG nova.network.neutron [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updating instance_info_cache with network_info: [{"id": "fc9835ce-e30c-4b58-8804-1062f5d166de", "address": "fa:16:3e:b8:59:0b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc9835ce-e3", "ovs_interfaceid": "fc9835ce-e30c-4b58-8804-1062f5d166de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7ce89428-70fc-4bff-a80f-235a2dc0a2a0", "address": "fa:16:3e:1f:65:b9", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ce89428-70", "ovs_interfaceid": "7ce89428-70fc-4bff-a80f-235a2dc0a2a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.789559] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.972115] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.153s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.972507] env[61906]: DEBUG nova.compute.manager [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1027.975176] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.186s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.976573] env[61906]: INFO nova.compute.claims [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1028.271573] env[61906]: DEBUG oslo_concurrency.lockutils [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.272373] env[61906]: DEBUG oslo_concurrency.lockutils [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.272575] env[61906]: DEBUG oslo_concurrency.lockutils [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.273534] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c14a297-3e75-401a-9a6c-6cff75d14f2d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.291484] env[61906]: DEBUG nova.virt.hardware [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1028.291741] env[61906]: DEBUG nova.virt.hardware [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1028.291943] env[61906]: DEBUG nova.virt.hardware [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1028.292212] env[61906]: DEBUG nova.virt.hardware [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1028.292406] env[61906]: DEBUG nova.virt.hardware [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1028.292593] env[61906]: DEBUG nova.virt.hardware [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1028.292846] env[61906]: DEBUG nova.virt.hardware [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1028.293062] env[61906]: DEBUG nova.virt.hardware [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1028.293274] env[61906]: DEBUG nova.virt.hardware [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1028.293470] env[61906]: DEBUG nova.virt.hardware [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1028.293679] env[61906]: DEBUG nova.virt.hardware [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1028.300412] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Reconfiguring VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1028.300733] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acc16923-eddc-43c7-ad54-749f73e700fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.318311] env[61906]: DEBUG oslo_vmware.api [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1028.318311] env[61906]: value = "task-1357058" [ 1028.318311] env[61906]: _type = "Task" [ 1028.318311] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.325906] env[61906]: DEBUG oslo_vmware.api [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357058, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.482061] env[61906]: DEBUG nova.compute.utils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1028.484464] env[61906]: DEBUG nova.compute.manager [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1028.484636] env[61906]: DEBUG nova.network.neutron [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1028.527924] env[61906]: DEBUG nova.policy [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e28284ad5ce450597ef12c65c902e1c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ac4f2c9065c46068c7837eb85cd7e20', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1028.779152] env[61906]: DEBUG nova.network.neutron [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Successfully created port: 7dccaaf0-b0f8-42db-bc51-b25e684a6115 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1028.830449] env[61906]: DEBUG oslo_vmware.api [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357058, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.985436] env[61906]: DEBUG nova.compute.manager [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1029.071725] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c6571ec-a915-4aca-b927-1448660433c6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.078948] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5845b55-bfe3-4217-9a29-22248c1acd63 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.109558] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104aaa03-f935-4a63-b1b2-d0d42ca98cd9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.116139] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c96ab17-020a-44a7-8271-5bdfd63176c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.128436] env[61906]: DEBUG nova.compute.provider_tree [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.328644] env[61906]: DEBUG oslo_vmware.api [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357058, 'name': ReconfigVM_Task, 'duration_secs': 0.566636} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.329200] env[61906]: DEBUG oslo_concurrency.lockutils [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.329424] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Reconfigured VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1029.631799] env[61906]: DEBUG nova.scheduler.client.report [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.835317] env[61906]: DEBUG oslo_concurrency.lockutils [None req-024de856-b618-4c62-8331-eb2acebf9304 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-3ea24d58-ffbd-41d4-a010-978a395ad2fc-7ce89428-70fc-4bff-a80f-235a2dc0a2a0" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.096s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.998291] env[61906]: DEBUG nova.compute.manager [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1030.025030] env[61906]: DEBUG nova.virt.hardware [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.025030] env[61906]: DEBUG nova.virt.hardware [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.025030] env[61906]: DEBUG nova.virt.hardware [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.025030] env[61906]: DEBUG nova.virt.hardware [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.025263] env[61906]: DEBUG nova.virt.hardware [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.025263] env[61906]: DEBUG nova.virt.hardware [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.025428] env[61906]: DEBUG nova.virt.hardware [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.025591] env[61906]: DEBUG nova.virt.hardware [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.025759] env[61906]: DEBUG nova.virt.hardware [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.025925] env[61906]: DEBUG nova.virt.hardware [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.026119] env[61906]: DEBUG nova.virt.hardware [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.027019] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd11a8d-dec7-48bd-9320-5beee237e233 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.034990] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47abb7b-6489-4461-a877-e550f0138cba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.136408] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.161s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.136968] env[61906]: DEBUG nova.compute.manager [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1030.159691] env[61906]: DEBUG nova.compute.manager [req-39c89600-f858-46fa-810e-d0bc4c18aceb req-c4b428b4-37dc-434e-9acc-f61c6278b66b service nova] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Received event network-vif-plugged-7dccaaf0-b0f8-42db-bc51-b25e684a6115 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.160044] env[61906]: DEBUG oslo_concurrency.lockutils [req-39c89600-f858-46fa-810e-d0bc4c18aceb req-c4b428b4-37dc-434e-9acc-f61c6278b66b service nova] Acquiring lock "1e1a742f-824f-4118-82fb-cafce626e0ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.160293] env[61906]: DEBUG oslo_concurrency.lockutils [req-39c89600-f858-46fa-810e-d0bc4c18aceb req-c4b428b4-37dc-434e-9acc-f61c6278b66b service nova] Lock "1e1a742f-824f-4118-82fb-cafce626e0ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.160476] env[61906]: DEBUG oslo_concurrency.lockutils [req-39c89600-f858-46fa-810e-d0bc4c18aceb req-c4b428b4-37dc-434e-9acc-f61c6278b66b service nova] Lock "1e1a742f-824f-4118-82fb-cafce626e0ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.160621] env[61906]: DEBUG nova.compute.manager [req-39c89600-f858-46fa-810e-d0bc4c18aceb req-c4b428b4-37dc-434e-9acc-f61c6278b66b service nova] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] No waiting events found dispatching network-vif-plugged-7dccaaf0-b0f8-42db-bc51-b25e684a6115 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1030.160932] env[61906]: WARNING nova.compute.manager [req-39c89600-f858-46fa-810e-d0bc4c18aceb req-c4b428b4-37dc-434e-9acc-f61c6278b66b service nova] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Received unexpected event network-vif-plugged-7dccaaf0-b0f8-42db-bc51-b25e684a6115 for instance with vm_state building and task_state spawning. [ 1030.248325] env[61906]: DEBUG nova.network.neutron [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Successfully updated port: 7dccaaf0-b0f8-42db-bc51-b25e684a6115 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1030.642322] env[61906]: DEBUG nova.compute.utils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1030.643816] env[61906]: DEBUG nova.compute.manager [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1030.643941] env[61906]: DEBUG nova.network.neutron [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1030.705500] env[61906]: DEBUG nova.policy [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ba1658b4ade413b98834102c39a255c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c6414fa082f49c2a568a772a256a8e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1030.750596] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "refresh_cache-1e1a742f-824f-4118-82fb-cafce626e0ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.750743] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "refresh_cache-1e1a742f-824f-4118-82fb-cafce626e0ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.750888] env[61906]: DEBUG nova.network.neutron [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1031.046098] env[61906]: DEBUG nova.network.neutron [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Successfully created port: da2e07e3-7453-4d91-88f7-26247bd7ea9d {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1031.147580] env[61906]: DEBUG nova.compute.manager [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1031.283007] env[61906]: DEBUG nova.network.neutron [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1031.493637] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "interface-3ea24d58-ffbd-41d4-a010-978a395ad2fc-7ce89428-70fc-4bff-a80f-235a2dc0a2a0" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.493988] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-3ea24d58-ffbd-41d4-a010-978a395ad2fc-7ce89428-70fc-4bff-a80f-235a2dc0a2a0" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.608763] env[61906]: DEBUG nova.network.neutron [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Updating instance_info_cache with network_info: [{"id": "7dccaaf0-b0f8-42db-bc51-b25e684a6115", "address": "fa:16:3e:d7:3c:a2", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7dccaaf0-b0", "ovs_interfaceid": "7dccaaf0-b0f8-42db-bc51-b25e684a6115", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.996820] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.997190] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.997915] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1a0df9-52b3-456c-8243-c700aaf7a327 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.015497] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778fbd21-b301-40c8-a9ce-55a3e0d74649 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.039911] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Reconfiguring VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1032.040270] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2925e2f7-26c9-4d5e-ac95-024b15aa927a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.058563] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1032.058563] env[61906]: value = "task-1357059" [ 1032.058563] env[61906]: _type = "Task" [ 1032.058563] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.067381] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357059, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.111112] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "refresh_cache-1e1a742f-824f-4118-82fb-cafce626e0ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.111449] env[61906]: DEBUG nova.compute.manager [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Instance network_info: |[{"id": "7dccaaf0-b0f8-42db-bc51-b25e684a6115", "address": "fa:16:3e:d7:3c:a2", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7dccaaf0-b0", "ovs_interfaceid": "7dccaaf0-b0f8-42db-bc51-b25e684a6115", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1032.111866] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:3c:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1895250-76cc-41f7-b7f8-2e5679494607', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7dccaaf0-b0f8-42db-bc51-b25e684a6115', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1032.119100] env[61906]: DEBUG oslo.service.loopingcall [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.119322] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1032.119545] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-279da367-3885-460e-91af-0c157f6d6be3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.138932] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1032.138932] env[61906]: value = "task-1357060" [ 1032.138932] env[61906]: _type = "Task" [ 1032.138932] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.146247] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357060, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.157603] env[61906]: DEBUG nova.compute.manager [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1032.186693] env[61906]: DEBUG nova.virt.hardware [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1032.187198] env[61906]: DEBUG nova.virt.hardware [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1032.187405] env[61906]: DEBUG nova.virt.hardware [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1032.187734] env[61906]: DEBUG nova.virt.hardware [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1032.188016] env[61906]: DEBUG nova.virt.hardware [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1032.188255] env[61906]: DEBUG nova.virt.hardware [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1032.188496] env[61906]: DEBUG nova.virt.hardware [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1032.188658] env[61906]: DEBUG nova.virt.hardware [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1032.188839] env[61906]: DEBUG nova.virt.hardware [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1032.188994] env[61906]: DEBUG nova.virt.hardware [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1032.189210] env[61906]: DEBUG nova.virt.hardware [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1032.190127] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326fe56c-f50a-44bb-95d0-8ba109c2dcdd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.194748] env[61906]: DEBUG nova.compute.manager [req-3f313dd7-3eef-42e7-9eec-8e72571812e5 req-83c25edf-125b-4b5b-a9af-46aab8c8092a service nova] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Received event network-changed-7dccaaf0-b0f8-42db-bc51-b25e684a6115 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.194941] env[61906]: DEBUG nova.compute.manager [req-3f313dd7-3eef-42e7-9eec-8e72571812e5 req-83c25edf-125b-4b5b-a9af-46aab8c8092a service nova] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Refreshing instance network info cache due to event network-changed-7dccaaf0-b0f8-42db-bc51-b25e684a6115. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1032.195178] env[61906]: DEBUG oslo_concurrency.lockutils [req-3f313dd7-3eef-42e7-9eec-8e72571812e5 req-83c25edf-125b-4b5b-a9af-46aab8c8092a service nova] Acquiring lock "refresh_cache-1e1a742f-824f-4118-82fb-cafce626e0ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.195329] env[61906]: DEBUG oslo_concurrency.lockutils [req-3f313dd7-3eef-42e7-9eec-8e72571812e5 req-83c25edf-125b-4b5b-a9af-46aab8c8092a service nova] Acquired lock "refresh_cache-1e1a742f-824f-4118-82fb-cafce626e0ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.195494] env[61906]: DEBUG nova.network.neutron [req-3f313dd7-3eef-42e7-9eec-8e72571812e5 req-83c25edf-125b-4b5b-a9af-46aab8c8092a service nova] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Refreshing network info cache for port 7dccaaf0-b0f8-42db-bc51-b25e684a6115 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1032.201755] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f10654a-c4a1-48f1-9fab-d2f1c8203826 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.495229] env[61906]: DEBUG nova.network.neutron [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Successfully updated port: da2e07e3-7453-4d91-88f7-26247bd7ea9d {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1032.568723] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.648709] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357060, 'name': CreateVM_Task, 'duration_secs': 0.319081} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.648948] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1032.649559] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.649733] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.650237] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1032.650512] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97bb845d-3108-46ca-8cc0-b79bbe0d5fb6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.654909] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1032.654909] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ce5bb4-9020-d28b-3fbe-98a37c592acb" [ 1032.654909] env[61906]: _type = "Task" [ 1032.654909] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.661991] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ce5bb4-9020-d28b-3fbe-98a37c592acb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.876441] env[61906]: DEBUG nova.network.neutron [req-3f313dd7-3eef-42e7-9eec-8e72571812e5 req-83c25edf-125b-4b5b-a9af-46aab8c8092a service nova] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Updated VIF entry in instance network info cache for port 7dccaaf0-b0f8-42db-bc51-b25e684a6115. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1032.876809] env[61906]: DEBUG nova.network.neutron [req-3f313dd7-3eef-42e7-9eec-8e72571812e5 req-83c25edf-125b-4b5b-a9af-46aab8c8092a service nova] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Updating instance_info_cache with network_info: [{"id": "7dccaaf0-b0f8-42db-bc51-b25e684a6115", "address": "fa:16:3e:d7:3c:a2", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7dccaaf0-b0", "ovs_interfaceid": "7dccaaf0-b0f8-42db-bc51-b25e684a6115", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.998217] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.998457] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.998547] env[61906]: DEBUG nova.network.neutron [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1033.068713] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.164895] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ce5bb4-9020-d28b-3fbe-98a37c592acb, 'name': SearchDatastore_Task, 'duration_secs': 0.010007} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.165222] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.165460] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1033.165697] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.165851] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.166052] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1033.166327] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-59d76e4e-099b-45ef-b37f-b462d5d176c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.174259] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1033.174441] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1033.175140] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-072d3b1b-0f69-48eb-80d3-ffffc9756e76 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.179703] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1033.179703] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52978d08-58ad-8582-6e4f-52195fa6bf21" [ 1033.179703] env[61906]: _type = "Task" [ 1033.179703] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.189320] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52978d08-58ad-8582-6e4f-52195fa6bf21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.379309] env[61906]: DEBUG oslo_concurrency.lockutils [req-3f313dd7-3eef-42e7-9eec-8e72571812e5 req-83c25edf-125b-4b5b-a9af-46aab8c8092a service nova] Releasing lock "refresh_cache-1e1a742f-824f-4118-82fb-cafce626e0ac" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.529063] env[61906]: DEBUG nova.network.neutron [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1033.571861] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.650711] env[61906]: DEBUG nova.network.neutron [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance_info_cache with network_info: [{"id": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "address": "fa:16:3e:c2:38:09", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda2e07e3-74", "ovs_interfaceid": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.690682] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52978d08-58ad-8582-6e4f-52195fa6bf21, 'name': SearchDatastore_Task, 'duration_secs': 0.00779} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.691616] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52f0c55f-11bc-4c25-b429-caed3aa6382a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.696670] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1033.696670] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]523b006b-7489-62ce-169d-d7e1450bd98c" [ 1033.696670] env[61906]: _type = "Task" [ 1033.696670] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.704323] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]523b006b-7489-62ce-169d-d7e1450bd98c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.069893] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.153749] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.154140] env[61906]: DEBUG nova.compute.manager [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Instance network_info: |[{"id": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "address": "fa:16:3e:c2:38:09", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda2e07e3-74", "ovs_interfaceid": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1034.154596] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:38:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8233b50c-be21-441a-a357-a29138a95b8b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'da2e07e3-7453-4d91-88f7-26247bd7ea9d', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1034.161892] env[61906]: DEBUG oslo.service.loopingcall [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.162151] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1034.162383] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-795e350a-58be-4788-b231-4308b2a3e012 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.182694] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1034.182694] env[61906]: value = "task-1357061" [ 1034.182694] env[61906]: _type = "Task" [ 1034.182694] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.189872] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357061, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.205271] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]523b006b-7489-62ce-169d-d7e1450bd98c, 'name': SearchDatastore_Task, 'duration_secs': 0.009339} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.205564] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.205939] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 1e1a742f-824f-4118-82fb-cafce626e0ac/1e1a742f-824f-4118-82fb-cafce626e0ac.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1034.206123] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e47db724-92a6-4442-ba89-154d519a1254 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.212437] env[61906]: DEBUG nova.compute.manager [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Received event network-vif-plugged-da2e07e3-7453-4d91-88f7-26247bd7ea9d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1034.212652] env[61906]: DEBUG oslo_concurrency.lockutils [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] Acquiring lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.212843] env[61906]: DEBUG oslo_concurrency.lockutils [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] Lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.213107] env[61906]: DEBUG oslo_concurrency.lockutils [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] Lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.213217] env[61906]: DEBUG nova.compute.manager [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] No waiting events found dispatching network-vif-plugged-da2e07e3-7453-4d91-88f7-26247bd7ea9d {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1034.213387] env[61906]: WARNING nova.compute.manager [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Received unexpected event network-vif-plugged-da2e07e3-7453-4d91-88f7-26247bd7ea9d for instance with vm_state building and task_state spawning. [ 1034.213548] env[61906]: DEBUG nova.compute.manager [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Received event network-changed-da2e07e3-7453-4d91-88f7-26247bd7ea9d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1034.213702] env[61906]: DEBUG nova.compute.manager [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Refreshing instance network info cache due to event network-changed-da2e07e3-7453-4d91-88f7-26247bd7ea9d. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1034.213958] env[61906]: DEBUG oslo_concurrency.lockutils [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] Acquiring lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.214130] env[61906]: DEBUG oslo_concurrency.lockutils [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] Acquired lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.214293] env[61906]: DEBUG nova.network.neutron [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Refreshing network info cache for port da2e07e3-7453-4d91-88f7-26247bd7ea9d {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1034.216919] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1034.216919] env[61906]: value = "task-1357062" [ 1034.216919] env[61906]: _type = "Task" [ 1034.216919] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.225884] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357062, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.571482] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.692339] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357061, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.727867] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357062, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488582} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.728170] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 1e1a742f-824f-4118-82fb-cafce626e0ac/1e1a742f-824f-4118-82fb-cafce626e0ac.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1034.728448] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1034.728714] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-260e7978-83df-4f41-9ecd-79cb0ae7da62 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.735123] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1034.735123] env[61906]: value = "task-1357063" [ 1034.735123] env[61906]: _type = "Task" [ 1034.735123] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.745075] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357063, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.070794] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.142533] env[61906]: DEBUG nova.network.neutron [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updated VIF entry in instance network info cache for port da2e07e3-7453-4d91-88f7-26247bd7ea9d. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1035.143057] env[61906]: DEBUG nova.network.neutron [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance_info_cache with network_info: [{"id": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "address": "fa:16:3e:c2:38:09", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda2e07e3-74", "ovs_interfaceid": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.194421] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357061, 'name': CreateVM_Task, 'duration_secs': 0.515058} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.194628] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1035.195333] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.195512] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.195835] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1035.196126] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6faef3f8-75c8-4277-bf1d-34d77a1b6515 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.200931] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1035.200931] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529ac1a2-6663-ac7a-4502-2e681a5d54f9" [ 1035.200931] env[61906]: _type = "Task" [ 1035.200931] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.208580] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529ac1a2-6663-ac7a-4502-2e681a5d54f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.244135] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357063, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077317} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.244393] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1035.245134] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fda0ec-7644-40ea-a244-ad2641fa98de {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.266245] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 1e1a742f-824f-4118-82fb-cafce626e0ac/1e1a742f-824f-4118-82fb-cafce626e0ac.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1035.266487] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c74a708b-15d1-43be-a8d7-571f5dcf60b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.284772] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1035.284772] env[61906]: value = "task-1357064" [ 1035.284772] env[61906]: _type = "Task" [ 1035.284772] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.292468] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357064, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.571733] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.645640] env[61906]: DEBUG oslo_concurrency.lockutils [req-359b4bf1-dd9f-459b-b104-63f425bb5beb req-5595d995-6c81-499f-a412-965b82d5cdaa service nova] Releasing lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.710927] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529ac1a2-6663-ac7a-4502-2e681a5d54f9, 'name': SearchDatastore_Task, 'duration_secs': 0.06808} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.711263] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.711504] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1035.711744] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.711898] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.712112] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1035.712381] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5385c46d-c9fb-4341-b241-ea293b81681b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.719678] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1035.719883] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1035.720552] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d10240f4-1ffc-45d1-83a4-14ef78ceb5cc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.724955] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1035.724955] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d4a35a-6614-c9b2-41c6-93899b52830b" [ 1035.724955] env[61906]: _type = "Task" [ 1035.724955] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.732100] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d4a35a-6614-c9b2-41c6-93899b52830b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.793285] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357064, 'name': ReconfigVM_Task, 'duration_secs': 0.260478} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.793554] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 1e1a742f-824f-4118-82fb-cafce626e0ac/1e1a742f-824f-4118-82fb-cafce626e0ac.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1035.794162] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4212e074-0a11-4ae4-b464-01bc4899951a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.800204] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1035.800204] env[61906]: value = "task-1357065" [ 1035.800204] env[61906]: _type = "Task" [ 1035.800204] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.807261] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357065, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.071418] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.234683] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d4a35a-6614-c9b2-41c6-93899b52830b, 'name': SearchDatastore_Task, 'duration_secs': 0.008066} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.235497] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-985743fa-2ce6-4f5d-82d6-f904fadd58e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.240944] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1036.240944] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529cebdb-e987-50c9-645c-86fb6d1e8f2d" [ 1036.240944] env[61906]: _type = "Task" [ 1036.240944] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.249724] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529cebdb-e987-50c9-645c-86fb6d1e8f2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.309466] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357065, 'name': Rename_Task, 'duration_secs': 0.119289} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.309776] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1036.310039] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2978a1e7-a47e-4b09-b558-02046b079ceb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.316342] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1036.316342] env[61906]: value = "task-1357066" [ 1036.316342] env[61906]: _type = "Task" [ 1036.316342] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.323387] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357066, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.573097] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.752067] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]529cebdb-e987-50c9-645c-86fb6d1e8f2d, 'name': SearchDatastore_Task, 'duration_secs': 0.008958} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.752397] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.752693] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] fc0e81b8-86b8-4396-afdb-fe9184815fc0/fc0e81b8-86b8-4396-afdb-fe9184815fc0.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1036.752989] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7183a66-1790-4c30-bcf2-9fef91dc82e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.759584] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1036.759584] env[61906]: value = "task-1357067" [ 1036.759584] env[61906]: _type = "Task" [ 1036.759584] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.766655] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357067, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.825901] env[61906]: DEBUG oslo_vmware.api [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357066, 'name': PowerOnVM_Task, 'duration_secs': 0.419372} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.826237] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1036.826456] env[61906]: INFO nova.compute.manager [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Took 6.83 seconds to spawn the instance on the hypervisor. [ 1036.826643] env[61906]: DEBUG nova.compute.manager [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1036.827404] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5798b8b1-771d-4228-b05a-3acff17aa7c2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.074553] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.272389] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357067, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.345847] env[61906]: INFO nova.compute.manager [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Took 11.54 seconds to build instance. [ 1037.573763] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357059, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.770812] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357067, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517456} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.771164] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] fc0e81b8-86b8-4396-afdb-fe9184815fc0/fc0e81b8-86b8-4396-afdb-fe9184815fc0.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1037.771455] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1037.771761] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-524bd4a3-9696-4811-81fa-30a0b524ed8d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.779326] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1037.779326] env[61906]: value = "task-1357068" [ 1037.779326] env[61906]: _type = "Task" [ 1037.779326] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.787530] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357068, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.848301] env[61906]: DEBUG oslo_concurrency.lockutils [None req-e1ac659e-3fcb-4221-974d-b5b7b6216412 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "1e1a742f-824f-4118-82fb-cafce626e0ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.053s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.076056] env[61906]: DEBUG oslo_vmware.api [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357059, 'name': ReconfigVM_Task, 'duration_secs': 5.765382} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.076368] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.076527] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Reconfigured VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1038.101089] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "1e1a742f-824f-4118-82fb-cafce626e0ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.101343] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "1e1a742f-824f-4118-82fb-cafce626e0ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.101551] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "1e1a742f-824f-4118-82fb-cafce626e0ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.101738] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "1e1a742f-824f-4118-82fb-cafce626e0ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.101908] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "1e1a742f-824f-4118-82fb-cafce626e0ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.105596] env[61906]: INFO nova.compute.manager [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Terminating instance [ 1038.107473] env[61906]: DEBUG nova.compute.manager [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1038.107675] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1038.108459] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359e3cd0-10fb-46ac-8f36-3d8ff1e33362 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.115213] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1038.115433] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5682ba46-8ffa-4348-aa4e-c9bb8722ce25 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.121773] env[61906]: DEBUG oslo_vmware.api [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1038.121773] env[61906]: value = "task-1357069" [ 1038.121773] env[61906]: _type = "Task" [ 1038.121773] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.129191] env[61906]: DEBUG oslo_vmware.api [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357069, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.290273] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357068, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.157234} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.290567] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1038.291389] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31c8097-fd38-4ae0-a6c3-eb2c809d6a87 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.312685] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] fc0e81b8-86b8-4396-afdb-fe9184815fc0/fc0e81b8-86b8-4396-afdb-fe9184815fc0.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1038.312975] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78babbe0-c231-458a-8ad2-56fd486f31b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.332565] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1038.332565] env[61906]: value = "task-1357070" [ 1038.332565] env[61906]: _type = "Task" [ 1038.332565] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.340553] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357070, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.631673] env[61906]: DEBUG oslo_vmware.api [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357069, 'name': PowerOffVM_Task, 'duration_secs': 0.485208} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.631673] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1038.631941] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1038.631941] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5c84a31-db5d-4f9d-a87d-6a64ece2edd9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.720091] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1038.720366] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1038.720561] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleting the datastore file [datastore2] 1e1a742f-824f-4118-82fb-cafce626e0ac {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.720862] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d6109303-1407-42aa-be61-89dbc20053b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.726669] env[61906]: DEBUG oslo_vmware.api [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1038.726669] env[61906]: value = "task-1357072" [ 1038.726669] env[61906]: _type = "Task" [ 1038.726669] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.734296] env[61906]: DEBUG oslo_vmware.api [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357072, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.842205] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357070, 'name': ReconfigVM_Task, 'duration_secs': 0.27447} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.842562] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Reconfigured VM instance instance-00000065 to attach disk [datastore2] fc0e81b8-86b8-4396-afdb-fe9184815fc0/fc0e81b8-86b8-4396-afdb-fe9184815fc0.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1038.843293] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8fca474c-8c3f-4220-938b-3e5a52bf312f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.849774] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1038.849774] env[61906]: value = "task-1357073" [ 1038.849774] env[61906]: _type = "Task" [ 1038.849774] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.857199] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357073, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.236131] env[61906]: DEBUG oslo_vmware.api [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357072, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16401} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.236441] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.236690] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1039.236911] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1039.237148] env[61906]: INFO nova.compute.manager [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1039.237446] env[61906]: DEBUG oslo.service.loopingcall [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.237678] env[61906]: DEBUG nova.compute.manager [-] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1039.237820] env[61906]: DEBUG nova.network.neutron [-] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1039.359229] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357073, 'name': Rename_Task, 'duration_secs': 0.136419} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.359533] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1039.359814] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0766c89c-6cc8-43a5-89e0-a66c88576d0c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.365328] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1039.365328] env[61906]: value = "task-1357074" [ 1039.365328] env[61906]: _type = "Task" [ 1039.365328] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.373798] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357074, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.386447] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.386643] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.386832] env[61906]: DEBUG nova.network.neutron [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1039.484499] env[61906]: DEBUG nova.compute.manager [req-fc599964-6414-4aa4-9386-0f2d39c0e1fc req-a66fce1e-5598-443e-95fb-b94754ea3308 service nova] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Received event network-vif-deleted-7dccaaf0-b0f8-42db-bc51-b25e684a6115 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.484760] env[61906]: INFO nova.compute.manager [req-fc599964-6414-4aa4-9386-0f2d39c0e1fc req-a66fce1e-5598-443e-95fb-b94754ea3308 service nova] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Neutron deleted interface 7dccaaf0-b0f8-42db-bc51-b25e684a6115; detaching it from the instance and deleting it from the info cache [ 1039.484911] env[61906]: DEBUG nova.network.neutron [req-fc599964-6414-4aa4-9386-0f2d39c0e1fc req-a66fce1e-5598-443e-95fb-b94754ea3308 service nova] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.877805] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357074, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.960414] env[61906]: DEBUG nova.network.neutron [-] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.988605] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-74459806-0890-439a-b198-f32637f5f594 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.997348] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532ac75e-bd3c-429f-b848-60ca983a1eb1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.028647] env[61906]: DEBUG nova.compute.manager [req-fc599964-6414-4aa4-9386-0f2d39c0e1fc req-a66fce1e-5598-443e-95fb-b94754ea3308 service nova] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Detach interface failed, port_id=7dccaaf0-b0f8-42db-bc51-b25e684a6115, reason: Instance 1e1a742f-824f-4118-82fb-cafce626e0ac could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1040.236631] env[61906]: INFO nova.network.neutron [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Port 7ce89428-70fc-4bff-a80f-235a2dc0a2a0 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1040.237177] env[61906]: DEBUG nova.network.neutron [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updating instance_info_cache with network_info: [{"id": "fc9835ce-e30c-4b58-8804-1062f5d166de", "address": "fa:16:3e:b8:59:0b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc9835ce-e3", "ovs_interfaceid": "fc9835ce-e30c-4b58-8804-1062f5d166de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.376473] env[61906]: DEBUG oslo_vmware.api [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357074, 'name': PowerOnVM_Task, 'duration_secs': 0.624809} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.376749] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1040.376959] env[61906]: INFO nova.compute.manager [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Took 8.22 seconds to spawn the instance on the hypervisor. [ 1040.377163] env[61906]: DEBUG nova.compute.manager [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1040.377915] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4fa71f0-ba4d-4252-ba40-795983b380b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.462985] env[61906]: INFO nova.compute.manager [-] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Took 1.22 seconds to deallocate network for instance. [ 1040.740059] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.899039] env[61906]: INFO nova.compute.manager [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Took 13.12 seconds to build instance. [ 1040.970330] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.970616] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.970864] env[61906]: DEBUG nova.objects.instance [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lazy-loading 'resources' on Instance uuid 1e1a742f-824f-4118-82fb-cafce626e0ac {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.979484] env[61906]: DEBUG oslo_concurrency.lockutils [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "interface-729f5565-8cda-4bee-8644-fb94a06034fb-7ce89428-70fc-4bff-a80f-235a2dc0a2a0" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.979847] env[61906]: DEBUG oslo_concurrency.lockutils [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-729f5565-8cda-4bee-8644-fb94a06034fb-7ce89428-70fc-4bff-a80f-235a2dc0a2a0" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.980382] env[61906]: DEBUG nova.objects.instance [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'flavor' on Instance uuid 729f5565-8cda-4bee-8644-fb94a06034fb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.247463] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d29bbaf6-8b69-4cc5-95b4-4cb469eab856 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-3ea24d58-ffbd-41d4-a010-978a395ad2fc-7ce89428-70fc-4bff-a80f-235a2dc0a2a0" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.753s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.263953] env[61906]: DEBUG nova.compute.manager [req-87b52f6b-7a60-4b64-b009-4407dabffda3 req-31fa54a2-850e-470f-8060-85da2990cb51 service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Received event network-changed-da2e07e3-7453-4d91-88f7-26247bd7ea9d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.263953] env[61906]: DEBUG nova.compute.manager [req-87b52f6b-7a60-4b64-b009-4407dabffda3 req-31fa54a2-850e-470f-8060-85da2990cb51 service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Refreshing instance network info cache due to event network-changed-da2e07e3-7453-4d91-88f7-26247bd7ea9d. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1041.263953] env[61906]: DEBUG oslo_concurrency.lockutils [req-87b52f6b-7a60-4b64-b009-4407dabffda3 req-31fa54a2-850e-470f-8060-85da2990cb51 service nova] Acquiring lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.263953] env[61906]: DEBUG oslo_concurrency.lockutils [req-87b52f6b-7a60-4b64-b009-4407dabffda3 req-31fa54a2-850e-470f-8060-85da2990cb51 service nova] Acquired lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.263953] env[61906]: DEBUG nova.network.neutron [req-87b52f6b-7a60-4b64-b009-4407dabffda3 req-31fa54a2-850e-470f-8060-85da2990cb51 service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Refreshing network info cache for port da2e07e3-7453-4d91-88f7-26247bd7ea9d {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1041.402679] env[61906]: DEBUG oslo_concurrency.lockutils [None req-88701b36-d94c-4c84-af17-cc64fbcb888c tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.635s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.516173] env[61906]: DEBUG nova.compute.manager [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Received event network-changed-fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.516388] env[61906]: DEBUG nova.compute.manager [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Refreshing instance network info cache due to event network-changed-fc9835ce-e30c-4b58-8804-1062f5d166de. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1041.516607] env[61906]: DEBUG oslo_concurrency.lockutils [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] Acquiring lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.516757] env[61906]: DEBUG oslo_concurrency.lockutils [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] Acquired lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.516927] env[61906]: DEBUG nova.network.neutron [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Refreshing network info cache for port fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1041.584018] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e3a6ae-c9cc-41af-94c8-4d798de83151 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.592169] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0efe8133-9bc8-4402-92c1-99b177f89ed7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.622348] env[61906]: DEBUG nova.objects.instance [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'pci_requests' on Instance uuid 729f5565-8cda-4bee-8644-fb94a06034fb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.623919] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f33d783-828a-475d-824a-25ff4be30343 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.632418] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2251d1-34da-4d60-bed4-8fb41cc064ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.648026] env[61906]: DEBUG nova.compute.provider_tree [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.988758] env[61906]: DEBUG nova.network.neutron [req-87b52f6b-7a60-4b64-b009-4407dabffda3 req-31fa54a2-850e-470f-8060-85da2990cb51 service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updated VIF entry in instance network info cache for port da2e07e3-7453-4d91-88f7-26247bd7ea9d. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1041.989338] env[61906]: DEBUG nova.network.neutron [req-87b52f6b-7a60-4b64-b009-4407dabffda3 req-31fa54a2-850e-470f-8060-85da2990cb51 service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance_info_cache with network_info: [{"id": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "address": "fa:16:3e:c2:38:09", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda2e07e3-74", "ovs_interfaceid": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.129024] env[61906]: DEBUG nova.objects.base [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Object Instance<729f5565-8cda-4bee-8644-fb94a06034fb> lazy-loaded attributes: flavor,pci_requests {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1042.129024] env[61906]: DEBUG nova.network.neutron [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1042.150751] env[61906]: DEBUG nova.scheduler.client.report [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1042.195479] env[61906]: DEBUG nova.policy [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73219f5e96f64d1da86a19da9b18b06a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5533be74874b4094b7ef0f98a121f60b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1042.224770] env[61906]: DEBUG nova.network.neutron [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updated VIF entry in instance network info cache for port fc9835ce-e30c-4b58-8804-1062f5d166de. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1042.225348] env[61906]: DEBUG nova.network.neutron [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updating instance_info_cache with network_info: [{"id": "fc9835ce-e30c-4b58-8804-1062f5d166de", "address": "fa:16:3e:b8:59:0b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc9835ce-e3", "ovs_interfaceid": "fc9835ce-e30c-4b58-8804-1062f5d166de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.491860] env[61906]: DEBUG oslo_concurrency.lockutils [req-87b52f6b-7a60-4b64-b009-4407dabffda3 req-31fa54a2-850e-470f-8060-85da2990cb51 service nova] Releasing lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.658171] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.687s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.683105] env[61906]: INFO nova.scheduler.client.report [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted allocations for instance 1e1a742f-824f-4118-82fb-cafce626e0ac [ 1042.728086] env[61906]: DEBUG oslo_concurrency.lockutils [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] Releasing lock "refresh_cache-3ea24d58-ffbd-41d4-a010-978a395ad2fc" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.728368] env[61906]: DEBUG nova.compute.manager [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Received event network-changed-a6bc2330-c2b4-4d0b-8178-f827316508a3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.728551] env[61906]: DEBUG nova.compute.manager [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Refreshing instance network info cache due to event network-changed-a6bc2330-c2b4-4d0b-8178-f827316508a3. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1042.728783] env[61906]: DEBUG oslo_concurrency.lockutils [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] Acquiring lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.728973] env[61906]: DEBUG oslo_concurrency.lockutils [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] Acquired lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.729164] env[61906]: DEBUG nova.network.neutron [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Refreshing network info cache for port a6bc2330-c2b4-4d0b-8178-f827316508a3 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1042.771221] env[61906]: DEBUG oslo_concurrency.lockutils [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.771480] env[61906]: DEBUG oslo_concurrency.lockutils [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.771669] env[61906]: INFO nova.compute.manager [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Shelving [ 1043.191691] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aeb71b18-17eb-495d-819b-aa52cd21d2c6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "1e1a742f-824f-4118-82fb-cafce626e0ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.090s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.194423] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1043.194674] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1043.194819] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Cleaning up deleted instances with incomplete migration {{(pid=61906) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1043.278130] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1043.278413] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d2590c9-d828-46fa-89df-bb0018d8ca94 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.286034] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1043.286034] env[61906]: value = "task-1357075" [ 1043.286034] env[61906]: _type = "Task" [ 1043.286034] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.294829] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357075, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.471235] env[61906]: DEBUG nova.network.neutron [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updated VIF entry in instance network info cache for port a6bc2330-c2b4-4d0b-8178-f827316508a3. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1043.471662] env[61906]: DEBUG nova.network.neutron [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updating instance_info_cache with network_info: [{"id": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "address": "fa:16:3e:87:df:8b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6bc2330-c2", "ovs_interfaceid": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.796218] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357075, 'name': PowerOffVM_Task, 'duration_secs': 0.23352} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.796531] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1043.797320] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-713c2576-f707-4f30-80d3-f2638fbe6b92 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.818888] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cddec5f6-4b1b-4ee1-b0fa-6cb00be295e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.902660] env[61906]: DEBUG nova.network.neutron [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Successfully updated port: 7ce89428-70fc-4bff-a80f-235a2dc0a2a0 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1043.945236] env[61906]: DEBUG nova.compute.manager [req-9ae64daf-416a-44da-bb77-14320c664938 req-3aee192c-3c6f-4262-8c1c-8ae2e156971c service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Received event network-vif-plugged-7ce89428-70fc-4bff-a80f-235a2dc0a2a0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1043.945484] env[61906]: DEBUG oslo_concurrency.lockutils [req-9ae64daf-416a-44da-bb77-14320c664938 req-3aee192c-3c6f-4262-8c1c-8ae2e156971c service nova] Acquiring lock "729f5565-8cda-4bee-8644-fb94a06034fb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.945704] env[61906]: DEBUG oslo_concurrency.lockutils [req-9ae64daf-416a-44da-bb77-14320c664938 req-3aee192c-3c6f-4262-8c1c-8ae2e156971c service nova] Lock "729f5565-8cda-4bee-8644-fb94a06034fb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.945898] env[61906]: DEBUG oslo_concurrency.lockutils [req-9ae64daf-416a-44da-bb77-14320c664938 req-3aee192c-3c6f-4262-8c1c-8ae2e156971c service nova] Lock "729f5565-8cda-4bee-8644-fb94a06034fb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.946084] env[61906]: DEBUG nova.compute.manager [req-9ae64daf-416a-44da-bb77-14320c664938 req-3aee192c-3c6f-4262-8c1c-8ae2e156971c service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] No waiting events found dispatching network-vif-plugged-7ce89428-70fc-4bff-a80f-235a2dc0a2a0 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1043.946259] env[61906]: WARNING nova.compute.manager [req-9ae64daf-416a-44da-bb77-14320c664938 req-3aee192c-3c6f-4262-8c1c-8ae2e156971c service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Received unexpected event network-vif-plugged-7ce89428-70fc-4bff-a80f-235a2dc0a2a0 for instance with vm_state active and task_state None. [ 1043.974751] env[61906]: DEBUG oslo_concurrency.lockutils [req-2119db11-0a9f-4a47-9361-76e520382c52 req-407f871f-0e05-41ac-9927-299c61ca16af service nova] Releasing lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.328771] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Creating Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1044.329122] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-053744b5-0cd3-4bb6-aa45-7b78be97a544 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.337013] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1044.337013] env[61906]: value = "task-1357076" [ 1044.337013] env[61906]: _type = "Task" [ 1044.337013] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.345043] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357076, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.402806] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "520e0d79-f7d9-4cc5-b480-a8f9e00f74ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.403055] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "520e0d79-f7d9-4cc5-b480-a8f9e00f74ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.407056] env[61906]: DEBUG oslo_concurrency.lockutils [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.407056] env[61906]: DEBUG oslo_concurrency.lockutils [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.407056] env[61906]: DEBUG nova.network.neutron [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1044.846567] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357076, 'name': CreateSnapshot_Task, 'duration_secs': 0.44634} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.846925] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Created Snapshot of the VM instance {{(pid=61906) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1044.847622] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-677775ad-ad2d-46da-9646-1ae27a0592af {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.905342] env[61906]: DEBUG nova.compute.manager [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1044.943207] env[61906]: WARNING nova.network.neutron [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] b12c63e1-bd61-4da3-8d37-4a89206ba0f2 already exists in list: networks containing: ['b12c63e1-bd61-4da3-8d37-4a89206ba0f2']. ignoring it [ 1045.260620] env[61906]: DEBUG nova.network.neutron [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updating instance_info_cache with network_info: [{"id": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "address": "fa:16:3e:87:df:8b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6bc2330-c2", "ovs_interfaceid": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7ce89428-70fc-4bff-a80f-235a2dc0a2a0", "address": "fa:16:3e:1f:65:b9", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ce89428-70", "ovs_interfaceid": "7ce89428-70fc-4bff-a80f-235a2dc0a2a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.364738] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Creating linked-clone VM from snapshot {{(pid=61906) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1045.365025] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-da0437d1-0c39-4f67-bc17-05354b5894fa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.373358] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1045.373358] env[61906]: value = "task-1357077" [ 1045.373358] env[61906]: _type = "Task" [ 1045.373358] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.382652] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357077, 'name': CloneVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.428321] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.428579] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.430166] env[61906]: INFO nova.compute.claims [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1045.764028] env[61906]: DEBUG oslo_concurrency.lockutils [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.764737] env[61906]: DEBUG oslo_concurrency.lockutils [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.764913] env[61906]: DEBUG oslo_concurrency.lockutils [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.765819] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3bfd4a-8417-47b3-9146-1aa0ea93670e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.786939] env[61906]: DEBUG nova.virt.hardware [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1045.787317] env[61906]: DEBUG nova.virt.hardware [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1045.787549] env[61906]: DEBUG nova.virt.hardware [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1045.787832] env[61906]: DEBUG nova.virt.hardware [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1045.788057] env[61906]: DEBUG nova.virt.hardware [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1045.788279] env[61906]: DEBUG nova.virt.hardware [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1045.788578] env[61906]: DEBUG nova.virt.hardware [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1045.788815] env[61906]: DEBUG nova.virt.hardware [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1045.789082] env[61906]: DEBUG nova.virt.hardware [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1045.789327] env[61906]: DEBUG nova.virt.hardware [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1045.789579] env[61906]: DEBUG nova.virt.hardware [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1045.799652] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Reconfiguring VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1045.800074] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9487cc2a-3f22-4e1c-bf8f-23113a64927b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.824656] env[61906]: DEBUG oslo_vmware.api [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1045.824656] env[61906]: value = "task-1357078" [ 1045.824656] env[61906]: _type = "Task" [ 1045.824656] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.832953] env[61906]: DEBUG oslo_vmware.api [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357078, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.883351] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357077, 'name': CloneVM_Task} progress is 94%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.976704] env[61906]: DEBUG nova.compute.manager [req-edd28c3c-56ed-496f-935e-08c41df4fc42 req-9c132622-eaf9-452a-a2eb-bc7f1200c46a service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Received event network-changed-7ce89428-70fc-4bff-a80f-235a2dc0a2a0 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.976904] env[61906]: DEBUG nova.compute.manager [req-edd28c3c-56ed-496f-935e-08c41df4fc42 req-9c132622-eaf9-452a-a2eb-bc7f1200c46a service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Refreshing instance network info cache due to event network-changed-7ce89428-70fc-4bff-a80f-235a2dc0a2a0. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1045.977149] env[61906]: DEBUG oslo_concurrency.lockutils [req-edd28c3c-56ed-496f-935e-08c41df4fc42 req-9c132622-eaf9-452a-a2eb-bc7f1200c46a service nova] Acquiring lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.977300] env[61906]: DEBUG oslo_concurrency.lockutils [req-edd28c3c-56ed-496f-935e-08c41df4fc42 req-9c132622-eaf9-452a-a2eb-bc7f1200c46a service nova] Acquired lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.977470] env[61906]: DEBUG nova.network.neutron [req-edd28c3c-56ed-496f-935e-08c41df4fc42 req-9c132622-eaf9-452a-a2eb-bc7f1200c46a service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Refreshing network info cache for port 7ce89428-70fc-4bff-a80f-235a2dc0a2a0 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1046.334497] env[61906]: DEBUG oslo_vmware.api [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357078, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.384009] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357077, 'name': CloneVM_Task} progress is 95%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.527496] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb5cfcc-6a06-4ca1-9ad3-32694535d7d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.535298] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578fe4a5-8d69-4ba7-865e-2a561eefc8d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.567171] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfdcef7e-4168-4862-89b0-5f91135bb207 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.574529] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9421ca26-21c0-48db-ab96-c9c6bd8d71d0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.587398] env[61906]: DEBUG nova.compute.provider_tree [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.697772] env[61906]: DEBUG nova.network.neutron [req-edd28c3c-56ed-496f-935e-08c41df4fc42 req-9c132622-eaf9-452a-a2eb-bc7f1200c46a service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updated VIF entry in instance network info cache for port 7ce89428-70fc-4bff-a80f-235a2dc0a2a0. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1046.698232] env[61906]: DEBUG nova.network.neutron [req-edd28c3c-56ed-496f-935e-08c41df4fc42 req-9c132622-eaf9-452a-a2eb-bc7f1200c46a service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updating instance_info_cache with network_info: [{"id": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "address": "fa:16:3e:87:df:8b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6bc2330-c2", "ovs_interfaceid": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7ce89428-70fc-4bff-a80f-235a2dc0a2a0", "address": "fa:16:3e:1f:65:b9", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ce89428-70", "ovs_interfaceid": "7ce89428-70fc-4bff-a80f-235a2dc0a2a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.834564] env[61906]: DEBUG oslo_vmware.api [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357078, 'name': ReconfigVM_Task, 'duration_secs': 0.610597} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.835091] env[61906]: DEBUG oslo_concurrency.lockutils [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.835324] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Reconfigured VM to attach interface {{(pid=61906) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1046.884305] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357077, 'name': CloneVM_Task, 'duration_secs': 1.120793} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.884624] env[61906]: INFO nova.virt.vmwareapi.vmops [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Created linked-clone VM from snapshot [ 1046.885274] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67e9aaf-55ad-46b4-9a0d-31a818a48084 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.892549] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Uploading image ce6e47d6-dc87-43bc-abae-811f9630f885 {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1046.918327] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1046.918327] env[61906]: value = "vm-289043" [ 1046.918327] env[61906]: _type = "VirtualMachine" [ 1046.918327] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1046.918987] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c45b64d9-5662-4af8-9e67-d707d71b4868 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.925619] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lease: (returnval){ [ 1046.925619] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b49da9-e450-23b7-27fa-53e30370d53d" [ 1046.925619] env[61906]: _type = "HttpNfcLease" [ 1046.925619] env[61906]: } obtained for exporting VM: (result){ [ 1046.925619] env[61906]: value = "vm-289043" [ 1046.925619] env[61906]: _type = "VirtualMachine" [ 1046.925619] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1046.925865] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the lease: (returnval){ [ 1046.925865] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b49da9-e450-23b7-27fa-53e30370d53d" [ 1046.925865] env[61906]: _type = "HttpNfcLease" [ 1046.925865] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1046.932469] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1046.932469] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b49da9-e450-23b7-27fa-53e30370d53d" [ 1046.932469] env[61906]: _type = "HttpNfcLease" [ 1046.932469] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1047.092117] env[61906]: DEBUG nova.scheduler.client.report [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.194537] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1047.194796] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1047.194955] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1047.195156] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1047.201026] env[61906]: DEBUG oslo_concurrency.lockutils [req-edd28c3c-56ed-496f-935e-08c41df4fc42 req-9c132622-eaf9-452a-a2eb-bc7f1200c46a service nova] Releasing lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.340083] env[61906]: DEBUG oslo_concurrency.lockutils [None req-233a6ac9-47da-4a88-b7fb-cf3103441d15 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-729f5565-8cda-4bee-8644-fb94a06034fb-7ce89428-70fc-4bff-a80f-235a2dc0a2a0" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.360s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.434181] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1047.434181] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b49da9-e450-23b7-27fa-53e30370d53d" [ 1047.434181] env[61906]: _type = "HttpNfcLease" [ 1047.434181] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1047.434427] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1047.434427] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52b49da9-e450-23b7-27fa-53e30370d53d" [ 1047.434427] env[61906]: _type = "HttpNfcLease" [ 1047.434427] env[61906]: }. {{(pid=61906) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1047.435145] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910f993c-1047-48c8-9403-28397877ec2b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.441827] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c36187-2f92-f903-d587-5cddf4996b53/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1047.442015] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c36187-2f92-f903-d587-5cddf4996b53/disk-0.vmdk for reading. {{(pid=61906) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1047.534878] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c765b03e-306b-46f5-95f0-c7d046830d54 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.597406] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.169s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.598030] env[61906]: DEBUG nova.compute.manager [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1048.103652] env[61906]: DEBUG nova.compute.utils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1048.105552] env[61906]: DEBUG nova.compute.manager [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1048.105811] env[61906]: DEBUG nova.network.neutron [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1048.164808] env[61906]: DEBUG nova.policy [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e28284ad5ce450597ef12c65c902e1c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ac4f2c9065c46068c7837eb85cd7e20', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1048.513704] env[61906]: DEBUG nova.network.neutron [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Successfully created port: e2b7bbde-7d23-403c-8b27-22bb05287ffb {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1048.609236] env[61906]: DEBUG nova.compute.manager [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1048.696450] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1048.696766] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1048.696925] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Cleaning up deleted instances {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1048.739935] env[61906]: DEBUG oslo_concurrency.lockutils [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "interface-729f5565-8cda-4bee-8644-fb94a06034fb-7ce89428-70fc-4bff-a80f-235a2dc0a2a0" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.740359] env[61906]: DEBUG oslo_concurrency.lockutils [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-729f5565-8cda-4bee-8644-fb94a06034fb-7ce89428-70fc-4bff-a80f-235a2dc0a2a0" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.208442] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] There are 36 instances to clean {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1049.208844] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 1e1a742f-824f-4118-82fb-cafce626e0ac] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1049.244255] env[61906]: DEBUG oslo_concurrency.lockutils [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.244458] env[61906]: DEBUG oslo_concurrency.lockutils [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.245355] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0247f7b-47c9-4746-b99c-e695ca49bd3a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.264572] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcef5b67-f55f-4bfc-ba79-2c2ce3f0bd4f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.295375] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Reconfiguring VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1049.296185] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad09eaf1-86ba-4afb-9450-aa03b9f16010 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.318207] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1049.318207] env[61906]: value = "task-1357080" [ 1049.318207] env[61906]: _type = "Task" [ 1049.318207] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.327479] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357080, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.620087] env[61906]: DEBUG nova.compute.manager [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1049.645446] env[61906]: DEBUG nova.virt.hardware [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1049.645718] env[61906]: DEBUG nova.virt.hardware [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1049.645884] env[61906]: DEBUG nova.virt.hardware [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1049.646084] env[61906]: DEBUG nova.virt.hardware [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1049.646239] env[61906]: DEBUG nova.virt.hardware [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1049.646394] env[61906]: DEBUG nova.virt.hardware [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1049.646602] env[61906]: DEBUG nova.virt.hardware [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1049.646766] env[61906]: DEBUG nova.virt.hardware [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1049.646943] env[61906]: DEBUG nova.virt.hardware [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1049.647122] env[61906]: DEBUG nova.virt.hardware [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1049.647306] env[61906]: DEBUG nova.virt.hardware [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1049.648183] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd776f0-a6f3-47b5-b4b0-c935fa62d4ce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.656411] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299e71bf-13d0-492f-81b8-bae7bf3ab804 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.712241] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: d2f16b79-c3fa-4450-8cb4-e9537215c60c] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1049.829215] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.898995] env[61906]: DEBUG nova.compute.manager [req-9b00598d-e2ec-4308-b539-bebae4bd11ed req-1a3e8b89-e167-4b73-a51f-81623d0fbbb2 service nova] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Received event network-vif-plugged-e2b7bbde-7d23-403c-8b27-22bb05287ffb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.899251] env[61906]: DEBUG oslo_concurrency.lockutils [req-9b00598d-e2ec-4308-b539-bebae4bd11ed req-1a3e8b89-e167-4b73-a51f-81623d0fbbb2 service nova] Acquiring lock "520e0d79-f7d9-4cc5-b480-a8f9e00f74ff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.899467] env[61906]: DEBUG oslo_concurrency.lockutils [req-9b00598d-e2ec-4308-b539-bebae4bd11ed req-1a3e8b89-e167-4b73-a51f-81623d0fbbb2 service nova] Lock "520e0d79-f7d9-4cc5-b480-a8f9e00f74ff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.899639] env[61906]: DEBUG oslo_concurrency.lockutils [req-9b00598d-e2ec-4308-b539-bebae4bd11ed req-1a3e8b89-e167-4b73-a51f-81623d0fbbb2 service nova] Lock "520e0d79-f7d9-4cc5-b480-a8f9e00f74ff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.899859] env[61906]: DEBUG nova.compute.manager [req-9b00598d-e2ec-4308-b539-bebae4bd11ed req-1a3e8b89-e167-4b73-a51f-81623d0fbbb2 service nova] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] No waiting events found dispatching network-vif-plugged-e2b7bbde-7d23-403c-8b27-22bb05287ffb {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1049.900074] env[61906]: WARNING nova.compute.manager [req-9b00598d-e2ec-4308-b539-bebae4bd11ed req-1a3e8b89-e167-4b73-a51f-81623d0fbbb2 service nova] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Received unexpected event network-vif-plugged-e2b7bbde-7d23-403c-8b27-22bb05287ffb for instance with vm_state building and task_state spawning. [ 1049.984833] env[61906]: DEBUG nova.network.neutron [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Successfully updated port: e2b7bbde-7d23-403c-8b27-22bb05287ffb {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1050.215272] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: e3578f5a-bd05-41fa-92e8-61d0e4ceed3f] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1050.329513] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.487774] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "refresh_cache-520e0d79-f7d9-4cc5-b480-a8f9e00f74ff" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.487915] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "refresh_cache-520e0d79-f7d9-4cc5-b480-a8f9e00f74ff" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.488154] env[61906]: DEBUG nova.network.neutron [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1050.718622] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 082b1a00-dd07-4d37-860c-6a91d6d40eaf] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1050.830356] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.020861] env[61906]: DEBUG nova.network.neutron [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1051.155729] env[61906]: DEBUG nova.network.neutron [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Updating instance_info_cache with network_info: [{"id": "e2b7bbde-7d23-403c-8b27-22bb05287ffb", "address": "fa:16:3e:99:8b:0e", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2b7bbde-7d", "ovs_interfaceid": "e2b7bbde-7d23-403c-8b27-22bb05287ffb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.223391] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 766e79cb-bea3-4b74-923d-a147c7be4134] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1051.330701] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.658836] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "refresh_cache-520e0d79-f7d9-4cc5-b480-a8f9e00f74ff" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.659228] env[61906]: DEBUG nova.compute.manager [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Instance network_info: |[{"id": "e2b7bbde-7d23-403c-8b27-22bb05287ffb", "address": "fa:16:3e:99:8b:0e", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2b7bbde-7d", "ovs_interfaceid": "e2b7bbde-7d23-403c-8b27-22bb05287ffb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1051.659729] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:8b:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1895250-76cc-41f7-b7f8-2e5679494607', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e2b7bbde-7d23-403c-8b27-22bb05287ffb', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1051.667688] env[61906]: DEBUG oslo.service.loopingcall [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1051.667973] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1051.668269] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d39c72af-4953-441f-bab4-1bca5dc6a8ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.689029] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1051.689029] env[61906]: value = "task-1357081" [ 1051.689029] env[61906]: _type = "Task" [ 1051.689029] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.696281] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357081, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.728110] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: c77823f1-189c-4ac9-9a40-7bc45e82b442] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1051.832352] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.924939] env[61906]: DEBUG nova.compute.manager [req-d9f733b2-2a9d-47cb-b5fb-f85f769a3756 req-9330d898-4083-4eaa-93a4-17455b8a1f02 service nova] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Received event network-changed-e2b7bbde-7d23-403c-8b27-22bb05287ffb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.925420] env[61906]: DEBUG nova.compute.manager [req-d9f733b2-2a9d-47cb-b5fb-f85f769a3756 req-9330d898-4083-4eaa-93a4-17455b8a1f02 service nova] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Refreshing instance network info cache due to event network-changed-e2b7bbde-7d23-403c-8b27-22bb05287ffb. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1051.925752] env[61906]: DEBUG oslo_concurrency.lockutils [req-d9f733b2-2a9d-47cb-b5fb-f85f769a3756 req-9330d898-4083-4eaa-93a4-17455b8a1f02 service nova] Acquiring lock "refresh_cache-520e0d79-f7d9-4cc5-b480-a8f9e00f74ff" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.925984] env[61906]: DEBUG oslo_concurrency.lockutils [req-d9f733b2-2a9d-47cb-b5fb-f85f769a3756 req-9330d898-4083-4eaa-93a4-17455b8a1f02 service nova] Acquired lock "refresh_cache-520e0d79-f7d9-4cc5-b480-a8f9e00f74ff" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.926212] env[61906]: DEBUG nova.network.neutron [req-d9f733b2-2a9d-47cb-b5fb-f85f769a3756 req-9330d898-4083-4eaa-93a4-17455b8a1f02 service nova] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Refreshing network info cache for port e2b7bbde-7d23-403c-8b27-22bb05287ffb {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1052.198955] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357081, 'name': CreateVM_Task, 'duration_secs': 0.34424} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.199174] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1052.199879] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.200109] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.200463] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1052.200722] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26db2ddb-03fb-4900-ad3b-c81be7732b35 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.206336] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1052.206336] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527f5818-451a-8876-aa82-40247ae43856" [ 1052.206336] env[61906]: _type = "Task" [ 1052.206336] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.214222] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527f5818-451a-8876-aa82-40247ae43856, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.231800] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 070f8882-9099-4f17-bae8-1701b5dcc08d] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1052.331026] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.620774] env[61906]: DEBUG nova.network.neutron [req-d9f733b2-2a9d-47cb-b5fb-f85f769a3756 req-9330d898-4083-4eaa-93a4-17455b8a1f02 service nova] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Updated VIF entry in instance network info cache for port e2b7bbde-7d23-403c-8b27-22bb05287ffb. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1052.621242] env[61906]: DEBUG nova.network.neutron [req-d9f733b2-2a9d-47cb-b5fb-f85f769a3756 req-9330d898-4083-4eaa-93a4-17455b8a1f02 service nova] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Updating instance_info_cache with network_info: [{"id": "e2b7bbde-7d23-403c-8b27-22bb05287ffb", "address": "fa:16:3e:99:8b:0e", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2b7bbde-7d", "ovs_interfaceid": "e2b7bbde-7d23-403c-8b27-22bb05287ffb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.717383] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]527f5818-451a-8876-aa82-40247ae43856, 'name': SearchDatastore_Task, 'duration_secs': 0.010606} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.717658] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.717934] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1052.718198] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.718357] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.718543] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1052.718817] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b3a93130-c828-4012-99a9-1c4b8a196360 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.726941] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1052.727139] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1052.727830] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-175e6ff7-e816-4e9e-9d84-5ebdbf46192f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.733104] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1052.733104] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52273613-cb83-9c80-4a2c-899adf44d2dd" [ 1052.733104] env[61906]: _type = "Task" [ 1052.733104] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.736801] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: a4c9f463-98e7-49eb-8c01-c8f0ee98a884] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1052.744398] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52273613-cb83-9c80-4a2c-899adf44d2dd, 'name': SearchDatastore_Task, 'duration_secs': 0.008598} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.745144] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2298724-0c22-4ad2-92f5-2d34ee07c520 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.749973] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1052.749973] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a5788c-d7be-ac47-4d8b-357231e2c025" [ 1052.749973] env[61906]: _type = "Task" [ 1052.749973] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.759491] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a5788c-d7be-ac47-4d8b-357231e2c025, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.833218] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.123949] env[61906]: DEBUG oslo_concurrency.lockutils [req-d9f733b2-2a9d-47cb-b5fb-f85f769a3756 req-9330d898-4083-4eaa-93a4-17455b8a1f02 service nova] Releasing lock "refresh_cache-520e0d79-f7d9-4cc5-b480-a8f9e00f74ff" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.240027] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: bc773559-fe14-485f-8bdc-1d4aab186521] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1053.261133] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a5788c-d7be-ac47-4d8b-357231e2c025, 'name': SearchDatastore_Task, 'duration_secs': 0.00793} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.261429] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.261728] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff/520e0d79-f7d9-4cc5-b480-a8f9e00f74ff.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1053.262051] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-573f9b57-613d-4036-9651-58d1759210df {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.271108] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1053.271108] env[61906]: value = "task-1357082" [ 1053.271108] env[61906]: _type = "Task" [ 1053.271108] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.279557] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357082, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.332860] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.744268] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: e35580b3-7177-46fd-bb86-b013efbf4911] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1053.782334] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357082, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.42251} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.782598] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff/520e0d79-f7d9-4cc5-b480-a8f9e00f74ff.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1053.782816] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1053.783088] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-05b514b4-af60-4966-94e9-31b2f80c7959 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.790140] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1053.790140] env[61906]: value = "task-1357083" [ 1053.790140] env[61906]: _type = "Task" [ 1053.790140] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.798138] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357083, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.833450] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.248324] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: d060a9f7-bd96-4e95-8780-1617a6ca7443] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1054.299557] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357083, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082204} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.299939] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1054.300743] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa7c83f-0b15-447a-be26-b4f0cc38ed14 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.323258] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff/520e0d79-f7d9-4cc5-b480-a8f9e00f74ff.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1054.323556] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ea78c4e-9c92-405f-9384-0b1a486cd43b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.346155] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357080, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.347542] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1054.347542] env[61906]: value = "task-1357084" [ 1054.347542] env[61906]: _type = "Task" [ 1054.347542] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.358152] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357084, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.752400] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 99e5c55d-79ae-47a6-8500-79fc68291650] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1054.847277] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357080, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.856547] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357084, 'name': ReconfigVM_Task, 'duration_secs': 0.270829} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.856838] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff/520e0d79-f7d9-4cc5-b480-a8f9e00f74ff.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1054.857494] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d17dcf0a-fd48-43d8-80f6-b8b1a6188279 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.863468] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1054.863468] env[61906]: value = "task-1357085" [ 1054.863468] env[61906]: _type = "Task" [ 1054.863468] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.871456] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357085, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.255869] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 03bd64e4-16fc-4659-a428-a2d9e7205b81] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1055.316799] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c36187-2f92-f903-d587-5cddf4996b53/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1055.317795] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8641184c-e412-424f-ba67-7865b6149fd8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.324447] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c36187-2f92-f903-d587-5cddf4996b53/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1055.324622] env[61906]: ERROR oslo_vmware.rw_handles [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c36187-2f92-f903-d587-5cddf4996b53/disk-0.vmdk due to incomplete transfer. [ 1055.324843] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2a6ff725-ce19-429e-8b56-996be811031e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.331121] env[61906]: DEBUG oslo_vmware.rw_handles [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c36187-2f92-f903-d587-5cddf4996b53/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1055.331322] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Uploaded image ce6e47d6-dc87-43bc-abae-811f9630f885 to the Glance image server {{(pid=61906) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1055.333422] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Destroying the VM {{(pid=61906) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1055.333648] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6c1eb63f-7b80-45ba-9d9d-977d34b6623f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.341869] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1055.341869] env[61906]: value = "task-1357086" [ 1055.341869] env[61906]: _type = "Task" [ 1055.341869] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.348380] env[61906]: DEBUG oslo_vmware.api [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357080, 'name': ReconfigVM_Task, 'duration_secs': 5.849457} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.348968] env[61906]: DEBUG oslo_concurrency.lockutils [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.349229] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Reconfigured VM to detach interface {{(pid=61906) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1055.354048] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357086, 'name': Destroy_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.371695] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357085, 'name': Rename_Task, 'duration_secs': 0.139662} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.371961] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1055.372201] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06d20827-4c34-451e-9294-23a25f7ced24 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.377960] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1055.377960] env[61906]: value = "task-1357087" [ 1055.377960] env[61906]: _type = "Task" [ 1055.377960] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.386894] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357087, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.760117] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: d0792521-4637-4ce7-a579-71caa80ba38c] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1055.851953] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357086, 'name': Destroy_Task, 'duration_secs': 0.323223} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.852293] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Destroyed the VM [ 1055.852553] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Deleting Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1055.852810] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-66fd4ff1-e184-400a-8723-210573400e31 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.858391] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1055.858391] env[61906]: value = "task-1357088" [ 1055.858391] env[61906]: _type = "Task" [ 1055.858391] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.865557] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357088, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.885794] env[61906]: DEBUG oslo_vmware.api [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357087, 'name': PowerOnVM_Task, 'duration_secs': 0.44414} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.886059] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1055.886267] env[61906]: INFO nova.compute.manager [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Took 6.27 seconds to spawn the instance on the hypervisor. [ 1055.886452] env[61906]: DEBUG nova.compute.manager [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1055.887190] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf39913b-6880-4840-8345-d102fea021f6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.264018] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: bb53f018-8546-4a00-a4be-d3c48416b1a1] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1056.369086] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357088, 'name': RemoveSnapshot_Task, 'duration_secs': 0.383497} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.369374] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Deleted Snapshot of the VM instance {{(pid=61906) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1056.369695] env[61906]: DEBUG nova.compute.manager [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1056.370473] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a16b26c-7ed4-40e6-b448-6b56a6ff0f63 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.403193] env[61906]: INFO nova.compute.manager [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Took 10.99 seconds to build instance. [ 1056.658519] env[61906]: DEBUG oslo_concurrency.lockutils [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.658733] env[61906]: DEBUG oslo_concurrency.lockutils [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquired lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.658911] env[61906]: DEBUG nova.network.neutron [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1056.768088] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 7c27acd5-d0ac-4eb0-8277-9a97ffcb7a63] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1056.881681] env[61906]: INFO nova.compute.manager [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Shelve offloading [ 1056.883306] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1056.883558] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48fc0921-bae1-4ba5-a39b-946ebe273e74 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.891210] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1056.891210] env[61906]: value = "task-1357089" [ 1056.891210] env[61906]: _type = "Task" [ 1056.891210] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.894363] env[61906]: DEBUG oslo_concurrency.lockutils [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "520e0d79-f7d9-4cc5-b480-a8f9e00f74ff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.899196] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357089, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.905667] env[61906]: DEBUG oslo_concurrency.lockutils [None req-bce664d4-06b9-4c5f-b816-d8d0b349af74 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "520e0d79-f7d9-4cc5-b480-a8f9e00f74ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.503s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.905929] env[61906]: DEBUG oslo_concurrency.lockutils [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "520e0d79-f7d9-4cc5-b480-a8f9e00f74ff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.012s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.906166] env[61906]: DEBUG oslo_concurrency.lockutils [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "520e0d79-f7d9-4cc5-b480-a8f9e00f74ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.906378] env[61906]: DEBUG oslo_concurrency.lockutils [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "520e0d79-f7d9-4cc5-b480-a8f9e00f74ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.906552] env[61906]: DEBUG oslo_concurrency.lockutils [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "520e0d79-f7d9-4cc5-b480-a8f9e00f74ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.908794] env[61906]: INFO nova.compute.manager [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Terminating instance [ 1056.910663] env[61906]: DEBUG nova.compute.manager [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1056.910861] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1056.911641] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec67ec0-c249-4f42-8dd3-361ef45c5e04 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.918529] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1056.918756] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b79601b4-9fd3-4f7c-b70c-4e2785ce86f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.924486] env[61906]: DEBUG oslo_vmware.api [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1056.924486] env[61906]: value = "task-1357090" [ 1056.924486] env[61906]: _type = "Task" [ 1056.924486] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.931666] env[61906]: DEBUG oslo_vmware.api [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357090, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.108290] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "729f5565-8cda-4bee-8644-fb94a06034fb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.108573] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "729f5565-8cda-4bee-8644-fb94a06034fb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.108849] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "729f5565-8cda-4bee-8644-fb94a06034fb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.109075] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "729f5565-8cda-4bee-8644-fb94a06034fb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.109261] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "729f5565-8cda-4bee-8644-fb94a06034fb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.111552] env[61906]: INFO nova.compute.manager [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Terminating instance [ 1057.113466] env[61906]: DEBUG nova.compute.manager [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1057.113662] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1057.114495] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a746b5d-3aab-4781-9cf3-3c0db7b316c6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.121486] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1057.121682] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e77c32c8-52cb-4424-861f-4fb13250e6d8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.127268] env[61906]: DEBUG oslo_vmware.api [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1057.127268] env[61906]: value = "task-1357091" [ 1057.127268] env[61906]: _type = "Task" [ 1057.127268] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.134750] env[61906]: DEBUG oslo_vmware.api [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357091, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.271953] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 3a6da3bd-8e28-4c23-8660-f32f5c862d02] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1057.404735] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] VM already powered off {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1057.404950] env[61906]: DEBUG nova.compute.manager [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1057.405739] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0de52ce6-3508-4e78-9fbd-7e3abc043dcd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.411878] env[61906]: DEBUG oslo_concurrency.lockutils [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.412110] env[61906]: DEBUG oslo_concurrency.lockutils [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.412300] env[61906]: DEBUG nova.network.neutron [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1057.430396] env[61906]: INFO nova.network.neutron [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Port 7ce89428-70fc-4bff-a80f-235a2dc0a2a0 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1057.430785] env[61906]: DEBUG nova.network.neutron [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updating instance_info_cache with network_info: [{"id": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "address": "fa:16:3e:87:df:8b", "network": {"id": "b12c63e1-bd61-4da3-8d37-4a89206ba0f2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1851156141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5533be74874b4094b7ef0f98a121f60b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6bc2330-c2", "ovs_interfaceid": "a6bc2330-c2b4-4d0b-8178-f827316508a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.434539] env[61906]: DEBUG oslo_vmware.api [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357090, 'name': PowerOffVM_Task, 'duration_secs': 0.169711} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.434775] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1057.434946] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1057.435757] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6ca9576f-3565-4673-81cc-fc8010a3e5d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.501669] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1057.502118] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1057.502372] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleting the datastore file [datastore1] 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1057.502764] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ecb4a9e-0e39-46db-a28d-b0df112e8e7f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.509058] env[61906]: DEBUG oslo_vmware.api [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1057.509058] env[61906]: value = "task-1357093" [ 1057.509058] env[61906]: _type = "Task" [ 1057.509058] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.516897] env[61906]: DEBUG oslo_vmware.api [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357093, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.637231] env[61906]: DEBUG oslo_vmware.api [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357091, 'name': PowerOffVM_Task, 'duration_secs': 0.185249} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.637584] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1057.637783] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1057.638064] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4607c1a7-cfd3-42d6-a502-0d13cf6ce59f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.699884] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1057.700138] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1057.700334] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Deleting the datastore file [datastore2] 729f5565-8cda-4bee-8644-fb94a06034fb {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1057.700604] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e693587-e3b2-4fb4-b376-3d4616e05c47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.706908] env[61906]: DEBUG oslo_vmware.api [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1057.706908] env[61906]: value = "task-1357095" [ 1057.706908] env[61906]: _type = "Task" [ 1057.706908] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.714812] env[61906]: DEBUG oslo_vmware.api [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357095, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.775900] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 63cbf664-078b-40c5-92bb-7a5a8d8aea3b] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1057.936282] env[61906]: DEBUG oslo_concurrency.lockutils [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Releasing lock "refresh_cache-729f5565-8cda-4bee-8644-fb94a06034fb" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.018657] env[61906]: DEBUG oslo_vmware.api [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357093, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191251} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.018912] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1058.019194] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1058.019399] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1058.019577] env[61906]: INFO nova.compute.manager [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1058.019852] env[61906]: DEBUG oslo.service.loopingcall [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.020062] env[61906]: DEBUG nova.compute.manager [-] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1058.020158] env[61906]: DEBUG nova.network.neutron [-] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1058.112229] env[61906]: DEBUG nova.network.neutron [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updating instance_info_cache with network_info: [{"id": "3c075a5e-3c44-45b5-bada-620921c0a005", "address": "fa:16:3e:39:60:50", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c075a5e-3c", "ovs_interfaceid": "3c075a5e-3c44-45b5-bada-620921c0a005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.219184] env[61906]: DEBUG oslo_vmware.api [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357095, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.189413} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.219184] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1058.219300] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1058.219709] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1058.219880] env[61906]: INFO nova.compute.manager [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1058.220270] env[61906]: DEBUG oslo.service.loopingcall [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.220563] env[61906]: DEBUG nova.compute.manager [-] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1058.220710] env[61906]: DEBUG nova.network.neutron [-] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1058.268758] env[61906]: DEBUG nova.compute.manager [req-33c64d9b-ba91-4a76-8307-594f72092020 req-6a00d92c-13a1-419e-966f-084b6e66b15c service nova] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Received event network-vif-deleted-e2b7bbde-7d23-403c-8b27-22bb05287ffb {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.268973] env[61906]: INFO nova.compute.manager [req-33c64d9b-ba91-4a76-8307-594f72092020 req-6a00d92c-13a1-419e-966f-084b6e66b15c service nova] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Neutron deleted interface e2b7bbde-7d23-403c-8b27-22bb05287ffb; detaching it from the instance and deleting it from the info cache [ 1058.269166] env[61906]: DEBUG nova.network.neutron [req-33c64d9b-ba91-4a76-8307-594f72092020 req-6a00d92c-13a1-419e-966f-084b6e66b15c service nova] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.280147] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 943c57f9-74c6-4465-9773-3fb01e78127b] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1058.440821] env[61906]: DEBUG oslo_concurrency.lockutils [None req-399ed404-3cb4-4632-8189-b94f4993f546 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "interface-729f5565-8cda-4bee-8644-fb94a06034fb-7ce89428-70fc-4bff-a80f-235a2dc0a2a0" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.700s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.614613] env[61906]: DEBUG oslo_concurrency.lockutils [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.736099] env[61906]: DEBUG nova.network.neutron [-] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.772186] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7a79924f-bffa-4ccd-a02c-c7b325e13d41 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.782186] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41068a6-4d40-4432-9d38-616882a37fa7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.792742] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 0af2b07b-0632-4376-b502-58fe23d79f02] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1058.809073] env[61906]: DEBUG nova.compute.manager [req-33c64d9b-ba91-4a76-8307-594f72092020 req-6a00d92c-13a1-419e-966f-084b6e66b15c service nova] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Detach interface failed, port_id=e2b7bbde-7d23-403c-8b27-22bb05287ffb, reason: Instance 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1058.971823] env[61906]: DEBUG nova.compute.manager [req-24b47c6b-b672-45a3-bcd8-c43cc78b82ea req-705189ab-ecc2-47f3-9997-ab0ee56a71bf service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Received event network-vif-deleted-a6bc2330-c2b4-4d0b-8178-f827316508a3 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.972053] env[61906]: INFO nova.compute.manager [req-24b47c6b-b672-45a3-bcd8-c43cc78b82ea req-705189ab-ecc2-47f3-9997-ab0ee56a71bf service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Neutron deleted interface a6bc2330-c2b4-4d0b-8178-f827316508a3; detaching it from the instance and deleting it from the info cache [ 1058.972237] env[61906]: DEBUG nova.network.neutron [req-24b47c6b-b672-45a3-bcd8-c43cc78b82ea req-705189ab-ecc2-47f3-9997-ab0ee56a71bf service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.981359] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1058.982397] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b3bf7a-2d48-4d46-87dd-57781e7e7cd5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.990769] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1058.991034] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03ca1a83-6602-41ad-9053-be1cfa458cbe {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.060713] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1059.060860] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1059.061773] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleting the datastore file [datastore2] d8c1fc04-d5ac-4c78-95e1-e135b22819c0 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.061773] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2430d671-61aa-462f-8b5c-842794c2af5f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.068035] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1059.068035] env[61906]: value = "task-1357097" [ 1059.068035] env[61906]: _type = "Task" [ 1059.068035] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.075491] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357097, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.239800] env[61906]: INFO nova.compute.manager [-] [instance: 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff] Took 1.22 seconds to deallocate network for instance. [ 1059.296380] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 31d0f640-47f4-4b72-b9fe-6ef4c76c6b44] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1059.438694] env[61906]: DEBUG nova.network.neutron [-] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.476087] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c0b964a-079e-4734-a4fe-788e96539acb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.485500] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4d0863-7b99-4607-939f-1981b1d93ec3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.510282] env[61906]: DEBUG nova.compute.manager [req-24b47c6b-b672-45a3-bcd8-c43cc78b82ea req-705189ab-ecc2-47f3-9997-ab0ee56a71bf service nova] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Detach interface failed, port_id=a6bc2330-c2b4-4d0b-8178-f827316508a3, reason: Instance 729f5565-8cda-4bee-8644-fb94a06034fb could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1059.577300] env[61906]: DEBUG oslo_vmware.api [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357097, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135295} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.577548] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1059.577877] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1059.578100] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1059.634693] env[61906]: INFO nova.scheduler.client.report [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleted allocations for instance d8c1fc04-d5ac-4c78-95e1-e135b22819c0 [ 1059.745623] env[61906]: DEBUG oslo_concurrency.lockutils [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.745908] env[61906]: DEBUG oslo_concurrency.lockutils [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.746158] env[61906]: DEBUG nova.objects.instance [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lazy-loading 'resources' on Instance uuid 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.800318] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: a5f780a2-0cb1-4da8-8276-82b9653bf6a7] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1059.941473] env[61906]: INFO nova.compute.manager [-] [instance: 729f5565-8cda-4bee-8644-fb94a06034fb] Took 1.72 seconds to deallocate network for instance. [ 1060.139398] env[61906]: DEBUG oslo_concurrency.lockutils [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.297094] env[61906]: DEBUG nova.compute.manager [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Received event network-vif-unplugged-3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1060.297464] env[61906]: DEBUG oslo_concurrency.lockutils [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] Acquiring lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.297464] env[61906]: DEBUG oslo_concurrency.lockutils [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.297760] env[61906]: DEBUG oslo_concurrency.lockutils [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.297793] env[61906]: DEBUG nova.compute.manager [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] No waiting events found dispatching network-vif-unplugged-3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1060.297963] env[61906]: WARNING nova.compute.manager [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Received unexpected event network-vif-unplugged-3c075a5e-3c44-45b5-bada-620921c0a005 for instance with vm_state shelved_offloaded and task_state None. [ 1060.298314] env[61906]: DEBUG nova.compute.manager [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Received event network-changed-3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1060.298692] env[61906]: DEBUG nova.compute.manager [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Refreshing instance network info cache due to event network-changed-3c075a5e-3c44-45b5-bada-620921c0a005. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1060.298692] env[61906]: DEBUG oslo_concurrency.lockutils [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] Acquiring lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.298905] env[61906]: DEBUG oslo_concurrency.lockutils [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] Acquired lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.299251] env[61906]: DEBUG nova.network.neutron [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Refreshing network info cache for port 3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1060.303478] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: cb879dc0-3af7-4279-aa28-66e2b8b4286d] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1060.325249] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c3bccb-11ff-439a-947c-91586b16e5e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.333224] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6012da9-fef6-4bac-bf22-391eb4aafe9a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.362497] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3a13eb-3d2a-4ab0-867e-695c216666f6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.370281] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2df5c15-e826-4393-b97d-5a579e218963 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.385144] env[61906]: DEBUG nova.compute.provider_tree [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.447543] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.806094] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 24b4a747-67aa-4388-aed6-cb646cd55765] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1060.888246] env[61906]: DEBUG nova.scheduler.client.report [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1061.006718] env[61906]: DEBUG nova.network.neutron [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updated VIF entry in instance network info cache for port 3c075a5e-3c44-45b5-bada-620921c0a005. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1061.007103] env[61906]: DEBUG nova.network.neutron [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updating instance_info_cache with network_info: [{"id": "3c075a5e-3c44-45b5-bada-620921c0a005", "address": "fa:16:3e:39:60:50", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap3c075a5e-3c", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.309309] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 0552e012-edcf-444c-a9b3-f1cea93b9e82] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1061.393813] env[61906]: DEBUG oslo_concurrency.lockutils [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.648s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.396062] env[61906]: DEBUG oslo_concurrency.lockutils [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.257s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.396327] env[61906]: DEBUG nova.objects.instance [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lazy-loading 'resources' on Instance uuid d8c1fc04-d5ac-4c78-95e1-e135b22819c0 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.414284] env[61906]: INFO nova.scheduler.client.report [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted allocations for instance 520e0d79-f7d9-4cc5-b480-a8f9e00f74ff [ 1061.510124] env[61906]: DEBUG oslo_concurrency.lockutils [req-c4753359-09ec-4ea8-af68-dbfaf85c72a3 req-018b3888-03f7-43b9-ac9a-cabb45583b32 service nova] Releasing lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.812568] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: e133605d-c630-4b9e-a314-bf496c853710] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1061.898786] env[61906]: DEBUG nova.objects.instance [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lazy-loading 'numa_topology' on Instance uuid d8c1fc04-d5ac-4c78-95e1-e135b22819c0 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.921886] env[61906]: DEBUG oslo_concurrency.lockutils [None req-376af10a-045b-4ad8-8360-5311a088a9a5 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "520e0d79-f7d9-4cc5-b480-a8f9e00f74ff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.016s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.316944] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 7bae658e-b050-4639-b34e-c2671ef5c773] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1062.401175] env[61906]: DEBUG nova.objects.base [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1062.477301] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9274d8-21dc-4e53-bb78-52fdb0479714 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.485143] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d91faa3e-f68a-494d-b97b-a9e263ef0c8f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.516485] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12013851-d39c-4523-86d9-c0c280ce7aec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.523728] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11183702-6ddb-48f3-8a24-90d35a6aeed0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.537796] env[61906]: DEBUG nova.compute.provider_tree [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.733826] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "f235d418-6dbb-41ac-8a7b-d959f31adc22" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.734103] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "f235d418-6dbb-41ac-8a7b-d959f31adc22" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.821212] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 79452791-59cb-4722-bb4a-8e59d8c4e641] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1062.859500] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.041654] env[61906]: DEBUG nova.scheduler.client.report [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1063.236553] env[61906]: DEBUG nova.compute.manager [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1063.324872] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 51fa46a9-b1b3-4115-b1d7-cc4f672f2e83] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1063.551333] env[61906]: DEBUG oslo_concurrency.lockutils [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.155s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.553730] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.106s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.553968] env[61906]: DEBUG nova.objects.instance [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'resources' on Instance uuid 729f5565-8cda-4bee-8644-fb94a06034fb {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.758583] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.828657] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 5adaa660-b736-4c11-9141-846cf475ccd5] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1064.064108] env[61906]: DEBUG oslo_concurrency.lockutils [None req-75e13bf2-6f25-4939-b75a-dfe6db8db81c tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.292s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.064463] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.205s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.065429] env[61906]: INFO nova.compute.manager [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Unshelving [ 1064.133954] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9f3c48-7334-4de5-b809-97a714164003 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.141380] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991abc8c-da6f-4694-985a-50982097d95a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.171663] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56af1ecc-0879-4d39-bea1-cc9d9fafdab5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.179029] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6fe2f95-478e-4ffc-ba2c-d28b23c1ea34 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.191779] env[61906]: DEBUG nova.compute.provider_tree [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.332186] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 9a23071d-40fd-4446-aa03-ecbddcafe3d6] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1064.694437] env[61906]: DEBUG nova.scheduler.client.report [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.835438] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 8ab6b473-d13c-4341-9789-992ac3aba6a2] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1065.084377] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.198972] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.645s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.201201] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.443s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.203062] env[61906]: INFO nova.compute.claims [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1065.221367] env[61906]: INFO nova.scheduler.client.report [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Deleted allocations for instance 729f5565-8cda-4bee-8644-fb94a06034fb [ 1065.338791] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 2ec250d6-94ae-4a99-a0bf-9bfc7b5318fc] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1065.727821] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b9ad74bc-0aef-4e9f-9792-b6e4b57dda00 tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "729f5565-8cda-4bee-8644-fb94a06034fb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.619s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.841851] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 7943ab1a-82df-48dd-874c-15dc015db51e] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1066.278500] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8781b3-9133-4570-890c-65f263376d57 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.285908] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a724addb-f1e8-4aa4-a513-ab9397abe874 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.314780] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1ac8dd-1868-468d-b791-c5fa716b6cd1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.322460] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba8be578-b7d7-4b30-9478-a90f26faec86 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.339292] env[61906]: DEBUG nova.compute.provider_tree [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.344862] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 77791983-cef1-4f07-9b62-d52d335b08c8] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1066.830334] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.830643] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.830882] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.831087] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.831292] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.833514] env[61906]: INFO nova.compute.manager [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Terminating instance [ 1066.835323] env[61906]: DEBUG nova.compute.manager [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1066.835551] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1066.836392] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b00632-2e57-4898-bf58-8b774904c1f5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.842283] env[61906]: DEBUG nova.scheduler.client.report [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1066.847543] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: a6e54432-a58e-49ba-a5c8-e6188f2ea4ff] Instance has had 0 of 5 cleanup attempts {{(pid=61906) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1066.849120] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1066.849504] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-38265836-820e-4558-bbe6-fe736ac36b27 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.856961] env[61906]: DEBUG oslo_vmware.api [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1066.856961] env[61906]: value = "task-1357098" [ 1066.856961] env[61906]: _type = "Task" [ 1066.856961] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.865293] env[61906]: DEBUG oslo_vmware.api [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357098, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.350612] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.149s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.351241] env[61906]: DEBUG nova.compute.manager [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1067.353899] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.270s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.354159] env[61906]: DEBUG nova.objects.instance [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lazy-loading 'pci_requests' on Instance uuid d8c1fc04-d5ac-4c78-95e1-e135b22819c0 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1067.366561] env[61906]: DEBUG oslo_vmware.api [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357098, 'name': PowerOffVM_Task, 'duration_secs': 0.171515} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.367338] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1067.367528] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1067.367774] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-062b2a74-910b-4a6d-9c24-f3b6622f3c2b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.432675] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1067.433291] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1067.433291] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Deleting the datastore file [datastore1] 3ea24d58-ffbd-41d4-a010-978a395ad2fc {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.433409] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-03b24c53-985b-46f3-b174-ae9c93aa64b2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.440096] env[61906]: DEBUG oslo_vmware.api [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for the task: (returnval){ [ 1067.440096] env[61906]: value = "task-1357100" [ 1067.440096] env[61906]: _type = "Task" [ 1067.440096] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.447013] env[61906]: DEBUG oslo_vmware.api [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357100, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.853480] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.856654] env[61906]: DEBUG nova.compute.utils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1067.858918] env[61906]: DEBUG nova.compute.manager [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1067.859023] env[61906]: DEBUG nova.network.neutron [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1067.862189] env[61906]: DEBUG nova.objects.instance [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lazy-loading 'numa_topology' on Instance uuid d8c1fc04-d5ac-4c78-95e1-e135b22819c0 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1067.898544] env[61906]: DEBUG nova.policy [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e28284ad5ce450597ef12c65c902e1c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ac4f2c9065c46068c7837eb85cd7e20', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1067.949262] env[61906]: DEBUG oslo_vmware.api [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Task: {'id': task-1357100, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139567} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.949667] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1067.949814] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1067.950073] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1067.950270] env[61906]: INFO nova.compute.manager [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1067.950520] env[61906]: DEBUG oslo.service.loopingcall [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.950960] env[61906]: DEBUG nova.compute.manager [-] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1067.950960] env[61906]: DEBUG nova.network.neutron [-] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1068.143589] env[61906]: DEBUG nova.network.neutron [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Successfully created port: 213895ba-b477-4936-a766-369ea1f71948 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1068.191803] env[61906]: DEBUG nova.compute.manager [req-cffbac09-3d9c-4c97-beef-037e0efbdf49 req-6b164dac-9f89-4161-8ee0-dc5fe4913c07 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Received event network-vif-deleted-fc9835ce-e30c-4b58-8804-1062f5d166de {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.192185] env[61906]: INFO nova.compute.manager [req-cffbac09-3d9c-4c97-beef-037e0efbdf49 req-6b164dac-9f89-4161-8ee0-dc5fe4913c07 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Neutron deleted interface fc9835ce-e30c-4b58-8804-1062f5d166de; detaching it from the instance and deleting it from the info cache [ 1068.192475] env[61906]: DEBUG nova.network.neutron [req-cffbac09-3d9c-4c97-beef-037e0efbdf49 req-6b164dac-9f89-4161-8ee0-dc5fe4913c07 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.358064] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1068.358454] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1068.359722] env[61906]: DEBUG nova.compute.manager [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1068.364305] env[61906]: INFO nova.compute.claims [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1068.663130] env[61906]: DEBUG nova.network.neutron [-] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.696440] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c8ffb65-1684-4cde-9687-93f616419de1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.706484] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6840c119-d3ec-49dd-810c-f7869154dedf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.729133] env[61906]: DEBUG nova.compute.manager [req-cffbac09-3d9c-4c97-beef-037e0efbdf49 req-6b164dac-9f89-4161-8ee0-dc5fe4913c07 service nova] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Detach interface failed, port_id=fc9835ce-e30c-4b58-8804-1062f5d166de, reason: Instance 3ea24d58-ffbd-41d4-a010-978a395ad2fc could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1069.165931] env[61906]: INFO nova.compute.manager [-] [instance: 3ea24d58-ffbd-41d4-a010-978a395ad2fc] Took 1.21 seconds to deallocate network for instance. [ 1069.369788] env[61906]: DEBUG nova.compute.manager [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1069.401199] env[61906]: DEBUG nova.virt.hardware [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1069.401481] env[61906]: DEBUG nova.virt.hardware [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1069.401653] env[61906]: DEBUG nova.virt.hardware [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1069.401849] env[61906]: DEBUG nova.virt.hardware [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1069.402008] env[61906]: DEBUG nova.virt.hardware [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1069.402170] env[61906]: DEBUG nova.virt.hardware [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1069.402383] env[61906]: DEBUG nova.virt.hardware [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1069.402550] env[61906]: DEBUG nova.virt.hardware [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1069.402789] env[61906]: DEBUG nova.virt.hardware [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1069.402906] env[61906]: DEBUG nova.virt.hardware [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1069.403098] env[61906]: DEBUG nova.virt.hardware [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1069.403978] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eedb575-eb0b-4dd9-a67e-a7210d62b9f0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.414708] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d6c61f3-8d76-46d6-8f38-22575425e25f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.452596] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b635820-8c05-40d5-9b38-1e8b9c03915c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.459612] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9a51f9-c4ad-4ab3-842d-737161d43883 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.488015] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2541ca49-82a0-4fad-ac96-508d22e05b34 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.495115] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b7d47f-efff-4770-aba7-a14522a7b669 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.508057] env[61906]: DEBUG nova.compute.provider_tree [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.537225] env[61906]: DEBUG nova.compute.manager [req-cdad67cc-3d54-408d-b4c3-0a5ed1743a95 req-c742b0ac-56d7-4214-addc-28234fa0cd52 service nova] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Received event network-vif-plugged-213895ba-b477-4936-a766-369ea1f71948 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.537503] env[61906]: DEBUG oslo_concurrency.lockutils [req-cdad67cc-3d54-408d-b4c3-0a5ed1743a95 req-c742b0ac-56d7-4214-addc-28234fa0cd52 service nova] Acquiring lock "f235d418-6dbb-41ac-8a7b-d959f31adc22-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.537729] env[61906]: DEBUG oslo_concurrency.lockutils [req-cdad67cc-3d54-408d-b4c3-0a5ed1743a95 req-c742b0ac-56d7-4214-addc-28234fa0cd52 service nova] Lock "f235d418-6dbb-41ac-8a7b-d959f31adc22-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.537955] env[61906]: DEBUG oslo_concurrency.lockutils [req-cdad67cc-3d54-408d-b4c3-0a5ed1743a95 req-c742b0ac-56d7-4214-addc-28234fa0cd52 service nova] Lock "f235d418-6dbb-41ac-8a7b-d959f31adc22-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.538175] env[61906]: DEBUG nova.compute.manager [req-cdad67cc-3d54-408d-b4c3-0a5ed1743a95 req-c742b0ac-56d7-4214-addc-28234fa0cd52 service nova] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] No waiting events found dispatching network-vif-plugged-213895ba-b477-4936-a766-369ea1f71948 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1069.538379] env[61906]: WARNING nova.compute.manager [req-cdad67cc-3d54-408d-b4c3-0a5ed1743a95 req-c742b0ac-56d7-4214-addc-28234fa0cd52 service nova] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Received unexpected event network-vif-plugged-213895ba-b477-4936-a766-369ea1f71948 for instance with vm_state building and task_state spawning. [ 1069.627388] env[61906]: DEBUG nova.network.neutron [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Successfully updated port: 213895ba-b477-4936-a766-369ea1f71948 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1069.674757] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.011387] env[61906]: DEBUG nova.scheduler.client.report [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1070.133203] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "refresh_cache-f235d418-6dbb-41ac-8a7b-d959f31adc22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.133379] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "refresh_cache-f235d418-6dbb-41ac-8a7b-d959f31adc22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.133532] env[61906]: DEBUG nova.network.neutron [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1070.516102] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.162s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.518305] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.846s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.518549] env[61906]: DEBUG nova.objects.instance [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lazy-loading 'resources' on Instance uuid 3ea24d58-ffbd-41d4-a010-978a395ad2fc {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.549957] env[61906]: INFO nova.network.neutron [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updating port 3c075a5e-3c44-45b5-bada-620921c0a005 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1070.662503] env[61906]: DEBUG nova.network.neutron [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1070.803126] env[61906]: DEBUG nova.network.neutron [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Updating instance_info_cache with network_info: [{"id": "213895ba-b477-4936-a766-369ea1f71948", "address": "fa:16:3e:1b:34:f2", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap213895ba-b4", "ovs_interfaceid": "213895ba-b477-4936-a766-369ea1f71948", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.092435] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3549d5-d10d-43ce-9e33-759eac9aa9ea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.100174] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a2dcf3-3781-4a73-8140-c973dc4830b5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.129942] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c66c9f-dfa2-4140-9504-bc40ccfee47a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.137524] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad858cc-aa77-4bf0-ae03-e2d9c6f9b887 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.151396] env[61906]: DEBUG nova.compute.provider_tree [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.305909] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "refresh_cache-f235d418-6dbb-41ac-8a7b-d959f31adc22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.306199] env[61906]: DEBUG nova.compute.manager [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Instance network_info: |[{"id": "213895ba-b477-4936-a766-369ea1f71948", "address": "fa:16:3e:1b:34:f2", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap213895ba-b4", "ovs_interfaceid": "213895ba-b477-4936-a766-369ea1f71948", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1071.306774] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:34:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1895250-76cc-41f7-b7f8-2e5679494607', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '213895ba-b477-4936-a766-369ea1f71948', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1071.314339] env[61906]: DEBUG oslo.service.loopingcall [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.314550] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1071.314779] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f01e6b2-0f4e-429c-8032-80eb8a05d375 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.334857] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1071.334857] env[61906]: value = "task-1357101" [ 1071.334857] env[61906]: _type = "Task" [ 1071.334857] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.343940] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357101, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.562569] env[61906]: DEBUG nova.compute.manager [req-a793bb47-b18e-47e6-971c-4780fda5eb56 req-d4cf6546-69ff-48e5-bc26-990a71c29093 service nova] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Received event network-changed-213895ba-b477-4936-a766-369ea1f71948 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1071.562808] env[61906]: DEBUG nova.compute.manager [req-a793bb47-b18e-47e6-971c-4780fda5eb56 req-d4cf6546-69ff-48e5-bc26-990a71c29093 service nova] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Refreshing instance network info cache due to event network-changed-213895ba-b477-4936-a766-369ea1f71948. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1071.563017] env[61906]: DEBUG oslo_concurrency.lockutils [req-a793bb47-b18e-47e6-971c-4780fda5eb56 req-d4cf6546-69ff-48e5-bc26-990a71c29093 service nova] Acquiring lock "refresh_cache-f235d418-6dbb-41ac-8a7b-d959f31adc22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.563214] env[61906]: DEBUG oslo_concurrency.lockutils [req-a793bb47-b18e-47e6-971c-4780fda5eb56 req-d4cf6546-69ff-48e5-bc26-990a71c29093 service nova] Acquired lock "refresh_cache-f235d418-6dbb-41ac-8a7b-d959f31adc22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.563386] env[61906]: DEBUG nova.network.neutron [req-a793bb47-b18e-47e6-971c-4780fda5eb56 req-d4cf6546-69ff-48e5-bc26-990a71c29093 service nova] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Refreshing network info cache for port 213895ba-b477-4936-a766-369ea1f71948 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1071.653962] env[61906]: DEBUG nova.scheduler.client.report [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.845101] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357101, 'name': CreateVM_Task, 'duration_secs': 0.29212} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.845290] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1071.845949] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.846143] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.846476] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1071.846732] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1e23f32-ae80-4fe8-958a-c553ba97bd0d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.851386] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1071.851386] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5292bcf9-a29d-f836-bb63-c19ef6e8b6be" [ 1071.851386] env[61906]: _type = "Task" [ 1071.851386] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.858587] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5292bcf9-a29d-f836-bb63-c19ef6e8b6be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.882722] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Didn't find any instances for network info cache update. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1071.882940] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.883117] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.883280] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.883436] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1072.158597] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.640s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.184664] env[61906]: INFO nova.scheduler.client.report [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Deleted allocations for instance 3ea24d58-ffbd-41d4-a010-978a395ad2fc [ 1072.275622] env[61906]: DEBUG nova.network.neutron [req-a793bb47-b18e-47e6-971c-4780fda5eb56 req-d4cf6546-69ff-48e5-bc26-990a71c29093 service nova] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Updated VIF entry in instance network info cache for port 213895ba-b477-4936-a766-369ea1f71948. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1072.276032] env[61906]: DEBUG nova.network.neutron [req-a793bb47-b18e-47e6-971c-4780fda5eb56 req-d4cf6546-69ff-48e5-bc26-990a71c29093 service nova] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Updating instance_info_cache with network_info: [{"id": "213895ba-b477-4936-a766-369ea1f71948", "address": "fa:16:3e:1b:34:f2", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap213895ba-b4", "ovs_interfaceid": "213895ba-b477-4936-a766-369ea1f71948", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.362359] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5292bcf9-a29d-f836-bb63-c19ef6e8b6be, 'name': SearchDatastore_Task, 'duration_secs': 0.009815} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.362655] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.362906] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1072.363169] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.363323] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.363511] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1072.363779] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dab3c5bc-ddf0-4ddb-b9a0-7d7eced0a983 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.371789] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1072.371998] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1072.372691] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-397bcbb0-ec31-4620-861c-6f0f6dff6037 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.375644] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.375644] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.375885] env[61906]: DEBUG nova.network.neutron [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1072.379502] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1072.379502] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5232d005-9a71-db66-87b6-f5ed5963611d" [ 1072.379502] env[61906]: _type = "Task" [ 1072.379502] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.385937] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.386166] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.386342] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.386494] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1072.389868] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601d5a69-fdfa-4ba1-8051-fd023f856c89 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.392485] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5232d005-9a71-db66-87b6-f5ed5963611d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.397852] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2283692-b8b0-46d2-a49e-9ac41a3f4023 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.412362] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44b0e7a-489d-44b0-b701-cafb491aaa22 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.418327] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121f912e-3dea-4d8f-ba49-d2fd37effce8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.446185] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181124MB free_disk=150GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1072.446388] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.446521] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.693081] env[61906]: DEBUG oslo_concurrency.lockutils [None req-c619b42d-168e-4b3d-a762-7a81c237006e tempest-AttachInterfacesTestJSON-1802675520 tempest-AttachInterfacesTestJSON-1802675520-project-member] Lock "3ea24d58-ffbd-41d4-a010-978a395ad2fc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.862s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.778393] env[61906]: DEBUG oslo_concurrency.lockutils [req-a793bb47-b18e-47e6-971c-4780fda5eb56 req-d4cf6546-69ff-48e5-bc26-990a71c29093 service nova] Releasing lock "refresh_cache-f235d418-6dbb-41ac-8a7b-d959f31adc22" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.890086] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5232d005-9a71-db66-87b6-f5ed5963611d, 'name': SearchDatastore_Task, 'duration_secs': 0.008708} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.890885] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e1c85df-b107-4f5a-a73b-d6ef7db6a241 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.895712] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1072.895712] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52249e7a-17d6-2a9d-d565-6780bee730fa" [ 1072.895712] env[61906]: _type = "Task" [ 1072.895712] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.903390] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52249e7a-17d6-2a9d-d565-6780bee730fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.281541] env[61906]: DEBUG nova.network.neutron [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updating instance_info_cache with network_info: [{"id": "3c075a5e-3c44-45b5-bada-620921c0a005", "address": "fa:16:3e:39:60:50", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c075a5e-3c", "ovs_interfaceid": "3c075a5e-3c44-45b5-bada-620921c0a005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.406269] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52249e7a-17d6-2a9d-d565-6780bee730fa, 'name': SearchDatastore_Task, 'duration_secs': 0.009826} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.406564] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.406832] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] f235d418-6dbb-41ac-8a7b-d959f31adc22/f235d418-6dbb-41ac-8a7b-d959f31adc22.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1073.407107] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d0053ba-6a22-4b54-a1e7-89e26f678ba9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.413293] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1073.413293] env[61906]: value = "task-1357102" [ 1073.413293] env[61906]: _type = "Task" [ 1073.413293] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.421747] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357102, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.472474] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 91dc442c-0272-419d-8ef9-06e279a3ea61 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1073.472759] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance fc0e81b8-86b8-4396-afdb-fe9184815fc0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1073.473060] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance d8c1fc04-d5ac-4c78-95e1-e135b22819c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1073.473221] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance f235d418-6dbb-41ac-8a7b-d959f31adc22 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1073.473521] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1073.473784] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1073.534646] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26e5f50-7ecf-4a37-91b0-67dd8ab1eeea {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.541694] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54cfd87-7f05-417e-95a6-8ab4c8aa9f2a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.575816] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65a1a96-1df5-44ec-ab08-b22427260841 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.584431] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc4c97e-03cd-4c91-a28e-883c68a303ad {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.599474] env[61906]: DEBUG nova.compute.provider_tree [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.602982] env[61906]: DEBUG nova.compute.manager [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Received event network-vif-plugged-3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1073.603301] env[61906]: DEBUG oslo_concurrency.lockutils [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] Acquiring lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.603583] env[61906]: DEBUG oslo_concurrency.lockutils [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.603816] env[61906]: DEBUG oslo_concurrency.lockutils [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.604205] env[61906]: DEBUG nova.compute.manager [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] No waiting events found dispatching network-vif-plugged-3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1073.604334] env[61906]: WARNING nova.compute.manager [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Received unexpected event network-vif-plugged-3c075a5e-3c44-45b5-bada-620921c0a005 for instance with vm_state shelved_offloaded and task_state spawning. [ 1073.604629] env[61906]: DEBUG nova.compute.manager [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Received event network-changed-3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1073.605132] env[61906]: DEBUG nova.compute.manager [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Refreshing instance network info cache due to event network-changed-3c075a5e-3c44-45b5-bada-620921c0a005. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1073.605393] env[61906]: DEBUG oslo_concurrency.lockutils [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] Acquiring lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.785223] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.788133] env[61906]: DEBUG oslo_concurrency.lockutils [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] Acquired lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.788133] env[61906]: DEBUG nova.network.neutron [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Refreshing network info cache for port 3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1073.813691] env[61906]: DEBUG nova.virt.hardware [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='afe6c4a40a66d776a8ccccc434bb2667',container_format='bare',created_at=2024-10-24T14:13:48Z,direct_url=,disk_format='vmdk',id=ce6e47d6-dc87-43bc-abae-811f9630f885,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-439821428-shelved',owner='8f3bb9ce83b3430bb9202d3c72e77b4c',properties=ImageMetaProps,protected=,size=31665664,status='active',tags=,updated_at=2024-10-24T14:14:01Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1073.813967] env[61906]: DEBUG nova.virt.hardware [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1073.814151] env[61906]: DEBUG nova.virt.hardware [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1073.814344] env[61906]: DEBUG nova.virt.hardware [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1073.814521] env[61906]: DEBUG nova.virt.hardware [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1073.814780] env[61906]: DEBUG nova.virt.hardware [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1073.815040] env[61906]: DEBUG nova.virt.hardware [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1073.815205] env[61906]: DEBUG nova.virt.hardware [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1073.815385] env[61906]: DEBUG nova.virt.hardware [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1073.815583] env[61906]: DEBUG nova.virt.hardware [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1073.815771] env[61906]: DEBUG nova.virt.hardware [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1073.817143] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb979624-61ac-469b-8431-80e4355011e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.825927] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c8f9bd-09a6-4574-b4e4-19f2ae6dc49d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.840948] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:60:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8edfde4-5a99-4745-956d-04da82ab1b85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c075a5e-3c44-45b5-bada-620921c0a005', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1073.848550] env[61906]: DEBUG oslo.service.loopingcall [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1073.848795] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1073.849033] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-805e56c3-1997-4dfb-89d1-8eefdab560c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.867805] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1073.867805] env[61906]: value = "task-1357103" [ 1073.867805] env[61906]: _type = "Task" [ 1073.867805] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.875109] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357103, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.924661] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357102, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.107520] env[61906]: DEBUG nova.scheduler.client.report [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1074.378023] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357103, 'name': CreateVM_Task, 'duration_secs': 0.410527} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.378227] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1074.378887] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ce6e47d6-dc87-43bc-abae-811f9630f885" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.379064] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ce6e47d6-dc87-43bc-abae-811f9630f885" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.379595] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/ce6e47d6-dc87-43bc-abae-811f9630f885" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1074.379717] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e984a29-d1c3-4c34-b159-16341b14d31d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.384270] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1074.384270] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c2f24d-2cb0-c58c-81eb-7197aee7fbfc" [ 1074.384270] env[61906]: _type = "Task" [ 1074.384270] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.392597] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c2f24d-2cb0-c58c-81eb-7197aee7fbfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.423264] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357102, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.556745] env[61906]: DEBUG nova.network.neutron [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updated VIF entry in instance network info cache for port 3c075a5e-3c44-45b5-bada-620921c0a005. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1074.557130] env[61906]: DEBUG nova.network.neutron [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updating instance_info_cache with network_info: [{"id": "3c075a5e-3c44-45b5-bada-620921c0a005", "address": "fa:16:3e:39:60:50", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c075a5e-3c", "ovs_interfaceid": "3c075a5e-3c44-45b5-bada-620921c0a005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.612803] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1074.613045] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.166s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.894247] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ce6e47d6-dc87-43bc-abae-811f9630f885" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.894556] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Processing image ce6e47d6-dc87-43bc-abae-811f9630f885 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1074.894738] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/ce6e47d6-dc87-43bc-abae-811f9630f885/ce6e47d6-dc87-43bc-abae-811f9630f885.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.894984] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquired lock "[datastore1] devstack-image-cache_base/ce6e47d6-dc87-43bc-abae-811f9630f885/ce6e47d6-dc87-43bc-abae-811f9630f885.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.895188] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1074.895437] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6adec5f5-bfc7-408e-8d9e-c50298b074ff {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.903221] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1074.903407] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1074.904161] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-751d7522-8170-47fb-9d2c-589ba85bccc0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.909285] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1074.909285] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ddab59-eb46-ad34-d66b-1ff0ecf45890" [ 1074.909285] env[61906]: _type = "Task" [ 1074.909285] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.916516] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ddab59-eb46-ad34-d66b-1ff0ecf45890, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.923578] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357102, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.448254} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.923812] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] f235d418-6dbb-41ac-8a7b-d959f31adc22/f235d418-6dbb-41ac-8a7b-d959f31adc22.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1074.924041] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1074.924291] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a664ed45-2806-44dd-9d80-97da21cd44e0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.930399] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1074.930399] env[61906]: value = "task-1357104" [ 1074.930399] env[61906]: _type = "Task" [ 1074.930399] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.937628] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357104, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.060636] env[61906]: DEBUG oslo_concurrency.lockutils [req-16a25e82-6ae8-4dcd-9395-6cea6d275890 req-9f452679-86a0-45f0-b792-accf348a6cfa service nova] Releasing lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.421512] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Preparing fetch location {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1075.421791] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Fetch image to [datastore1] OSTACK_IMG_00e352d9-f738-4794-bbe1-4a27d6a3c404/OSTACK_IMG_00e352d9-f738-4794-bbe1-4a27d6a3c404.vmdk {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1075.421988] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Downloading stream optimized image ce6e47d6-dc87-43bc-abae-811f9630f885 to [datastore1] OSTACK_IMG_00e352d9-f738-4794-bbe1-4a27d6a3c404/OSTACK_IMG_00e352d9-f738-4794-bbe1-4a27d6a3c404.vmdk on the data store datastore1 as vApp {{(pid=61906) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1075.422197] env[61906]: DEBUG nova.virt.vmwareapi.images [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Downloading image file data ce6e47d6-dc87-43bc-abae-811f9630f885 to the ESX as VM named 'OSTACK_IMG_00e352d9-f738-4794-bbe1-4a27d6a3c404' {{(pid=61906) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1075.439752] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357104, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071318} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.440048] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1075.440896] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101d3de8-fcfb-48f1-8339-752d7eac34d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.464030] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] f235d418-6dbb-41ac-8a7b-d959f31adc22/f235d418-6dbb-41ac-8a7b-d959f31adc22.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1075.480105] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-449b2cfd-7eec-4532-a9d0-ad06120c6999 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.501678] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1075.501678] env[61906]: value = "task-1357105" [ 1075.501678] env[61906]: _type = "Task" [ 1075.501678] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.510237] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357105, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.532181] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1075.532181] env[61906]: value = "resgroup-9" [ 1075.532181] env[61906]: _type = "ResourcePool" [ 1075.532181] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1075.532480] env[61906]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-c4eb36ad-32b6-41a3-8b20-aac749b15f47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.552536] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lease: (returnval){ [ 1075.552536] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ebaafe-d216-08e4-6ec2-0f14461811d0" [ 1075.552536] env[61906]: _type = "HttpNfcLease" [ 1075.552536] env[61906]: } obtained for vApp import into resource pool (val){ [ 1075.552536] env[61906]: value = "resgroup-9" [ 1075.552536] env[61906]: _type = "ResourcePool" [ 1075.552536] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1075.552800] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the lease: (returnval){ [ 1075.552800] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ebaafe-d216-08e4-6ec2-0f14461811d0" [ 1075.552800] env[61906]: _type = "HttpNfcLease" [ 1075.552800] env[61906]: } to be ready. {{(pid=61906) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1075.559612] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1075.559612] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ebaafe-d216-08e4-6ec2-0f14461811d0" [ 1075.559612] env[61906]: _type = "HttpNfcLease" [ 1075.559612] env[61906]: } is initializing. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1076.011497] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357105, 'name': ReconfigVM_Task, 'duration_secs': 0.283411} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.011815] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Reconfigured VM instance instance-00000067 to attach disk [datastore1] f235d418-6dbb-41ac-8a7b-d959f31adc22/f235d418-6dbb-41ac-8a7b-d959f31adc22.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1076.012482] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-56ba0333-215e-4145-8ec5-b5001f909ba1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.018932] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1076.018932] env[61906]: value = "task-1357107" [ 1076.018932] env[61906]: _type = "Task" [ 1076.018932] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.026111] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357107, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.060807] env[61906]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1076.060807] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ebaafe-d216-08e4-6ec2-0f14461811d0" [ 1076.060807] env[61906]: _type = "HttpNfcLease" [ 1076.060807] env[61906]: } is ready. {{(pid=61906) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1076.061136] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1076.061136] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ebaafe-d216-08e4-6ec2-0f14461811d0" [ 1076.061136] env[61906]: _type = "HttpNfcLease" [ 1076.061136] env[61906]: }. {{(pid=61906) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1076.061892] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69cb0a91-341c-489b-96c0-14bac8229384 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.069021] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207817d-9fdc-d7cf-7f35-8a61fde27ffd/disk-0.vmdk from lease info. {{(pid=61906) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1076.069256] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating HTTP connection to write to file with size = 31665664 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207817d-9fdc-d7cf-7f35-8a61fde27ffd/disk-0.vmdk. {{(pid=61906) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1076.134272] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fee23600-bcdd-493d-9d52-c6adc98ff74e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.530883] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357107, 'name': Rename_Task, 'duration_secs': 0.132775} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.531180] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1076.531510] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0fba726-0ba9-43e9-b63e-06ea65730209 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.539084] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1076.539084] env[61906]: value = "task-1357108" [ 1076.539084] env[61906]: _type = "Task" [ 1076.539084] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.548239] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357108, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.049172] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357108, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.227416] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Completed reading data from the image iterator. {{(pid=61906) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1077.227766] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207817d-9fdc-d7cf-7f35-8a61fde27ffd/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1077.228801] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-369f86b5-5947-48e5-bbab-f15f0e1893e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.235917] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207817d-9fdc-d7cf-7f35-8a61fde27ffd/disk-0.vmdk is in state: ready. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1077.236113] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207817d-9fdc-d7cf-7f35-8a61fde27ffd/disk-0.vmdk. {{(pid=61906) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1077.236385] env[61906]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-8bf488bf-5ce0-4be0-b11f-bcbea6667129 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.432368] env[61906]: DEBUG oslo_vmware.rw_handles [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207817d-9fdc-d7cf-7f35-8a61fde27ffd/disk-0.vmdk. {{(pid=61906) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1077.432614] env[61906]: INFO nova.virt.vmwareapi.images [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Downloaded image file data ce6e47d6-dc87-43bc-abae-811f9630f885 [ 1077.434918] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52fb52b0-aaab-464f-b413-6df1c361d92f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.449275] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3ad03f5f-c278-4123-a897-d22e313b26e6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.484803] env[61906]: INFO nova.virt.vmwareapi.images [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] The imported VM was unregistered [ 1077.488050] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Caching image {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1077.488383] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Creating directory with path [datastore1] devstack-image-cache_base/ce6e47d6-dc87-43bc-abae-811f9630f885 {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1077.488792] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ffc000d9-109b-478e-bc92-e9ee702e5681 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.512973] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Created directory with path [datastore1] devstack-image-cache_base/ce6e47d6-dc87-43bc-abae-811f9630f885 {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1077.512973] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_00e352d9-f738-4794-bbe1-4a27d6a3c404/OSTACK_IMG_00e352d9-f738-4794-bbe1-4a27d6a3c404.vmdk to [datastore1] devstack-image-cache_base/ce6e47d6-dc87-43bc-abae-811f9630f885/ce6e47d6-dc87-43bc-abae-811f9630f885.vmdk. {{(pid=61906) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1077.513788] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-6656652d-15cd-4edb-98cd-a7af1abba0d9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.520066] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1077.520066] env[61906]: value = "task-1357110" [ 1077.520066] env[61906]: _type = "Task" [ 1077.520066] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.528316] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357110, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.548653] env[61906]: DEBUG oslo_vmware.api [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357108, 'name': PowerOnVM_Task, 'duration_secs': 0.544889} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.548920] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1077.549139] env[61906]: INFO nova.compute.manager [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Took 8.18 seconds to spawn the instance on the hypervisor. [ 1077.549322] env[61906]: DEBUG nova.compute.manager [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1077.550125] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07f283a-dd03-4174-997b-673ca57491b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.030354] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357110, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.068140] env[61906]: INFO nova.compute.manager [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Took 14.33 seconds to build instance. [ 1078.362661] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquiring lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.362965] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.531755] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357110, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.570591] env[61906]: DEBUG oslo_concurrency.lockutils [None req-73dfd28b-b6a7-473c-b3d4-5d267f7d3a32 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "f235d418-6dbb-41ac-8a7b-d959f31adc22" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.836s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.715719] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "fdd638bd-00b0-40e7-92df-ba0a378a0ae6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.715968] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "fdd638bd-00b0-40e7-92df-ba0a378a0ae6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.867022] env[61906]: DEBUG nova.compute.manager [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1079.036260] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357110, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.218749] env[61906]: DEBUG nova.compute.manager [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1079.392459] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.393326] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.394412] env[61906]: INFO nova.compute.claims [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1079.532954] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357110, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.596019] env[61906]: DEBUG nova.compute.manager [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Stashing vm_state: active {{(pid=61906) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1079.741295] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.033989] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357110, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.115466] env[61906]: DEBUG oslo_concurrency.lockutils [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.495832] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e4f71c4-6010-40e0-822e-a4cd6d81ac7c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.503449] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7e9b0e-78f7-4a00-a283-5bf61fdd26ed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.537743] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf85ec3-b0d2-4bb4-82e2-c6bd76ae662b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.544755] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357110, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.836204} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.546739] env[61906]: INFO nova.virt.vmwareapi.ds_util [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_00e352d9-f738-4794-bbe1-4a27d6a3c404/OSTACK_IMG_00e352d9-f738-4794-bbe1-4a27d6a3c404.vmdk to [datastore1] devstack-image-cache_base/ce6e47d6-dc87-43bc-abae-811f9630f885/ce6e47d6-dc87-43bc-abae-811f9630f885.vmdk. [ 1080.546945] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Cleaning up location [datastore1] OSTACK_IMG_00e352d9-f738-4794-bbe1-4a27d6a3c404 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1080.547129] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_00e352d9-f738-4794-bbe1-4a27d6a3c404 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1080.547426] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c53319f4-09a0-4daa-a881-bf0ea348210f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.549830] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a1bc88-37e8-4346-b406-d58af2649a1a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.563915] env[61906]: DEBUG nova.compute.provider_tree [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.566247] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1080.566247] env[61906]: value = "task-1357111" [ 1080.566247] env[61906]: _type = "Task" [ 1080.566247] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.573398] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357111, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.068240] env[61906]: DEBUG nova.scheduler.client.report [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.080191] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357111, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.329436} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.080430] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1081.080600] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Releasing lock "[datastore1] devstack-image-cache_base/ce6e47d6-dc87-43bc-abae-811f9630f885/ce6e47d6-dc87-43bc-abae-811f9630f885.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.080888] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/ce6e47d6-dc87-43bc-abae-811f9630f885/ce6e47d6-dc87-43bc-abae-811f9630f885.vmdk to [datastore1] d8c1fc04-d5ac-4c78-95e1-e135b22819c0/d8c1fc04-d5ac-4c78-95e1-e135b22819c0.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1081.081163] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-18f783d3-94d4-4005-97a9-39bbae30c4f1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.086920] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1081.086920] env[61906]: value = "task-1357112" [ 1081.086920] env[61906]: _type = "Task" [ 1081.086920] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.095216] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357112, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.576344] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.183s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.576891] env[61906]: DEBUG nova.compute.manager [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1081.579966] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.839s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.581370] env[61906]: INFO nova.compute.claims [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1081.595787] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357112, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.086417] env[61906]: DEBUG nova.compute.utils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1082.090860] env[61906]: DEBUG nova.compute.manager [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1082.091059] env[61906]: DEBUG nova.network.neutron [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1082.105626] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357112, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.143285] env[61906]: DEBUG nova.policy [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae88e93238d140cf95e8dce38559ef7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7849d220ba9f4c37a7c0c73535c17301', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1082.439164] env[61906]: DEBUG nova.network.neutron [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Successfully created port: ef2c8975-d452-47be-8622-b0a062d3b4b1 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1082.591683] env[61906]: DEBUG nova.compute.manager [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1082.615183] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357112, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.710283] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a1420e-2c5f-4aa5-a8f1-618e933d4392 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.720458] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574e40bd-2132-4a07-9404-9e9e5b252d22 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.755286] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81532ee-cb64-42ef-b52d-41aa1e315340 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.764808] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21384de-3924-41f7-86a9-febe9b27c3fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.782293] env[61906]: DEBUG nova.compute.provider_tree [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.109900] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357112, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.285972] env[61906]: DEBUG nova.scheduler.client.report [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1083.605021] env[61906]: DEBUG nova.compute.manager [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1083.610809] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357112, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.634571] env[61906]: DEBUG nova.virt.hardware [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1083.634911] env[61906]: DEBUG nova.virt.hardware [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1083.635110] env[61906]: DEBUG nova.virt.hardware [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1083.635313] env[61906]: DEBUG nova.virt.hardware [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1083.635468] env[61906]: DEBUG nova.virt.hardware [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1083.635630] env[61906]: DEBUG nova.virt.hardware [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1083.635849] env[61906]: DEBUG nova.virt.hardware [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1083.636100] env[61906]: DEBUG nova.virt.hardware [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1083.636313] env[61906]: DEBUG nova.virt.hardware [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1083.636601] env[61906]: DEBUG nova.virt.hardware [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1083.636816] env[61906]: DEBUG nova.virt.hardware [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1083.637794] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9434511-1e86-4117-b9af-aec23dbe69ab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.648200] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf34ca9-1092-4d1a-b2aa-91c386520f46 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.792069] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.212s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.792692] env[61906]: DEBUG nova.compute.manager [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1083.795421] env[61906]: DEBUG oslo_concurrency.lockutils [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.680s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.869150] env[61906]: DEBUG nova.compute.manager [req-70567e02-ebc5-40cf-bfaf-0f547a62c286 req-da0b0e4b-b7dd-48ea-a53b-58371a5063d5 service nova] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Received event network-vif-plugged-ef2c8975-d452-47be-8622-b0a062d3b4b1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.869488] env[61906]: DEBUG oslo_concurrency.lockutils [req-70567e02-ebc5-40cf-bfaf-0f547a62c286 req-da0b0e4b-b7dd-48ea-a53b-58371a5063d5 service nova] Acquiring lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.869823] env[61906]: DEBUG oslo_concurrency.lockutils [req-70567e02-ebc5-40cf-bfaf-0f547a62c286 req-da0b0e4b-b7dd-48ea-a53b-58371a5063d5 service nova] Lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.870191] env[61906]: DEBUG oslo_concurrency.lockutils [req-70567e02-ebc5-40cf-bfaf-0f547a62c286 req-da0b0e4b-b7dd-48ea-a53b-58371a5063d5 service nova] Lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.870516] env[61906]: DEBUG nova.compute.manager [req-70567e02-ebc5-40cf-bfaf-0f547a62c286 req-da0b0e4b-b7dd-48ea-a53b-58371a5063d5 service nova] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] No waiting events found dispatching network-vif-plugged-ef2c8975-d452-47be-8622-b0a062d3b4b1 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1083.870826] env[61906]: WARNING nova.compute.manager [req-70567e02-ebc5-40cf-bfaf-0f547a62c286 req-da0b0e4b-b7dd-48ea-a53b-58371a5063d5 service nova] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Received unexpected event network-vif-plugged-ef2c8975-d452-47be-8622-b0a062d3b4b1 for instance with vm_state building and task_state spawning. [ 1083.954084] env[61906]: DEBUG nova.network.neutron [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Successfully updated port: ef2c8975-d452-47be-8622-b0a062d3b4b1 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1084.105741] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357112, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.862035} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.106010] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/ce6e47d6-dc87-43bc-abae-811f9630f885/ce6e47d6-dc87-43bc-abae-811f9630f885.vmdk to [datastore1] d8c1fc04-d5ac-4c78-95e1-e135b22819c0/d8c1fc04-d5ac-4c78-95e1-e135b22819c0.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1084.106792] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d9702d-2bf4-46d7-8b31-cf3fa259e2a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.127710] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] d8c1fc04-d5ac-4c78-95e1-e135b22819c0/d8c1fc04-d5ac-4c78-95e1-e135b22819c0.vmdk or device None with type streamOptimized {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1084.127957] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4506d5f9-fc79-4582-b6fe-6fb70917d5a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.147067] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1084.147067] env[61906]: value = "task-1357113" [ 1084.147067] env[61906]: _type = "Task" [ 1084.147067] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.154367] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357113, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.298476] env[61906]: DEBUG nova.compute.utils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1084.300854] env[61906]: DEBUG nova.compute.manager [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1084.300969] env[61906]: DEBUG nova.network.neutron [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1084.305950] env[61906]: INFO nova.compute.claims [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1084.354110] env[61906]: DEBUG nova.policy [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e28284ad5ce450597ef12c65c902e1c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ac4f2c9065c46068c7837eb85cd7e20', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1084.459348] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquiring lock "refresh_cache-1dae8ad9-245f-4b05-9c65-59c2e19af22b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.459348] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquired lock "refresh_cache-1dae8ad9-245f-4b05-9c65-59c2e19af22b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.459348] env[61906]: DEBUG nova.network.neutron [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1084.658603] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357113, 'name': ReconfigVM_Task, 'duration_secs': 0.300917} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.658603] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Reconfigured VM instance instance-00000061 to attach disk [datastore1] d8c1fc04-d5ac-4c78-95e1-e135b22819c0/d8c1fc04-d5ac-4c78-95e1-e135b22819c0.vmdk or device None with type streamOptimized {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.659346] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-24d6e3d6-bf17-4ea3-b3ee-05674d28213d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.662589] env[61906]: DEBUG nova.network.neutron [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Successfully created port: fb600c17-7915-4f83-982e-605f7b6f1f61 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1084.665956] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1084.665956] env[61906]: value = "task-1357114" [ 1084.665956] env[61906]: _type = "Task" [ 1084.665956] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.673807] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357114, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.810820] env[61906]: DEBUG nova.compute.manager [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1084.815868] env[61906]: INFO nova.compute.resource_tracker [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating resource usage from migration 06224cea-26d8-443a-b299-09b795928f14 [ 1084.918369] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462dfa65-6f77-46ae-8dd4-6db42e9e8f0b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.926441] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d365f3-e479-40b8-9312-2aa2cd37c0e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.958719] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec9c342-7134-449b-931b-c63babd47054 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.967697] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e2298f4-de1e-4f96-a503-1a929495405b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.980897] env[61906]: DEBUG nova.compute.provider_tree [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1085.175890] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357114, 'name': Rename_Task, 'duration_secs': 0.140467} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.176204] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1085.176499] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dcd98ecc-b6d9-4262-9966-0d0260e983b7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.182682] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1085.182682] env[61906]: value = "task-1357115" [ 1085.182682] env[61906]: _type = "Task" [ 1085.182682] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.186114] env[61906]: DEBUG nova.network.neutron [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1085.192728] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357115, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.370138] env[61906]: DEBUG nova.network.neutron [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Updating instance_info_cache with network_info: [{"id": "ef2c8975-d452-47be-8622-b0a062d3b4b1", "address": "fa:16:3e:f6:7c:45", "network": {"id": "aee89e54-f939-4aa4-9212-df451d432c3a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-702450817-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7849d220ba9f4c37a7c0c73535c17301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec1528b-3e87-477b-8ab2-02696ad47e66", "external-id": "nsx-vlan-transportzone-180", "segmentation_id": 180, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef2c8975-d4", "ovs_interfaceid": "ef2c8975-d452-47be-8622-b0a062d3b4b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.484513] env[61906]: DEBUG nova.scheduler.client.report [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1085.695091] env[61906]: DEBUG oslo_vmware.api [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357115, 'name': PowerOnVM_Task, 'duration_secs': 0.46109} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.695453] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1085.800934] env[61906]: DEBUG nova.compute.manager [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1085.801881] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e8146b-40d5-4e0a-9c49-b8038f3bd34c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.827373] env[61906]: DEBUG nova.compute.manager [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1085.847566] env[61906]: DEBUG nova.virt.hardware [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1085.847821] env[61906]: DEBUG nova.virt.hardware [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1085.847987] env[61906]: DEBUG nova.virt.hardware [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1085.848195] env[61906]: DEBUG nova.virt.hardware [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1085.848349] env[61906]: DEBUG nova.virt.hardware [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1085.848504] env[61906]: DEBUG nova.virt.hardware [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1085.848718] env[61906]: DEBUG nova.virt.hardware [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1085.848887] env[61906]: DEBUG nova.virt.hardware [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1085.849076] env[61906]: DEBUG nova.virt.hardware [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1085.849252] env[61906]: DEBUG nova.virt.hardware [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1085.849428] env[61906]: DEBUG nova.virt.hardware [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1085.850360] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b339be6-21b2-43c3-ad9b-760d05e040b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.859118] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2b940e-004c-4616-8c3a-13a49601ee94 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.873828] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Releasing lock "refresh_cache-1dae8ad9-245f-4b05-9c65-59c2e19af22b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.874147] env[61906]: DEBUG nova.compute.manager [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Instance network_info: |[{"id": "ef2c8975-d452-47be-8622-b0a062d3b4b1", "address": "fa:16:3e:f6:7c:45", "network": {"id": "aee89e54-f939-4aa4-9212-df451d432c3a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-702450817-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7849d220ba9f4c37a7c0c73535c17301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec1528b-3e87-477b-8ab2-02696ad47e66", "external-id": "nsx-vlan-transportzone-180", "segmentation_id": 180, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef2c8975-d4", "ovs_interfaceid": "ef2c8975-d452-47be-8622-b0a062d3b4b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1085.874690] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:7c:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec1528b-3e87-477b-8ab2-02696ad47e66', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef2c8975-d452-47be-8622-b0a062d3b4b1', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1085.881870] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Creating folder: Project (7849d220ba9f4c37a7c0c73535c17301). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1085.882110] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e598c935-29fd-43c4-9ff3-a9f0dd4e3197 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.892011] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Created folder: Project (7849d220ba9f4c37a7c0c73535c17301) in parent group-v288914. [ 1085.892163] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Creating folder: Instances. Parent ref: group-v289048. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1085.893266] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52197a4e-ab94-4b02-8c1a-5acd9cd6129a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.895835] env[61906]: DEBUG nova.compute.manager [req-435b8738-8e3e-4cbc-bb21-e19b3beee001 req-5aa4a4b2-92cf-4885-973d-aeab01f356d1 service nova] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Received event network-changed-ef2c8975-d452-47be-8622-b0a062d3b4b1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1085.896032] env[61906]: DEBUG nova.compute.manager [req-435b8738-8e3e-4cbc-bb21-e19b3beee001 req-5aa4a4b2-92cf-4885-973d-aeab01f356d1 service nova] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Refreshing instance network info cache due to event network-changed-ef2c8975-d452-47be-8622-b0a062d3b4b1. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1085.896259] env[61906]: DEBUG oslo_concurrency.lockutils [req-435b8738-8e3e-4cbc-bb21-e19b3beee001 req-5aa4a4b2-92cf-4885-973d-aeab01f356d1 service nova] Acquiring lock "refresh_cache-1dae8ad9-245f-4b05-9c65-59c2e19af22b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.896471] env[61906]: DEBUG oslo_concurrency.lockutils [req-435b8738-8e3e-4cbc-bb21-e19b3beee001 req-5aa4a4b2-92cf-4885-973d-aeab01f356d1 service nova] Acquired lock "refresh_cache-1dae8ad9-245f-4b05-9c65-59c2e19af22b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.896573] env[61906]: DEBUG nova.network.neutron [req-435b8738-8e3e-4cbc-bb21-e19b3beee001 req-5aa4a4b2-92cf-4885-973d-aeab01f356d1 service nova] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Refreshing network info cache for port ef2c8975-d452-47be-8622-b0a062d3b4b1 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1085.905392] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Created folder: Instances in parent group-v289048. [ 1085.905619] env[61906]: DEBUG oslo.service.loopingcall [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1085.905805] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1085.906014] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f76fe3d-bffe-46c2-b833-a4917113d4ae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.924460] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1085.924460] env[61906]: value = "task-1357118" [ 1085.924460] env[61906]: _type = "Task" [ 1085.924460] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.931671] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357118, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.989905] env[61906]: DEBUG oslo_concurrency.lockutils [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.194s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.990035] env[61906]: INFO nova.compute.manager [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Migrating [ 1086.120597] env[61906]: DEBUG nova.network.neutron [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Successfully updated port: fb600c17-7915-4f83-982e-605f7b6f1f61 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1086.318595] env[61906]: DEBUG oslo_concurrency.lockutils [None req-a80c391f-4609-4d5d-8994-54fecc4b8ab8 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 22.254s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.434448] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357118, 'name': CreateVM_Task} progress is 99%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.504121] env[61906]: DEBUG oslo_concurrency.lockutils [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.504367] env[61906]: DEBUG oslo_concurrency.lockutils [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.504623] env[61906]: DEBUG nova.network.neutron [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1086.603469] env[61906]: DEBUG nova.network.neutron [req-435b8738-8e3e-4cbc-bb21-e19b3beee001 req-5aa4a4b2-92cf-4885-973d-aeab01f356d1 service nova] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Updated VIF entry in instance network info cache for port ef2c8975-d452-47be-8622-b0a062d3b4b1. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1086.603940] env[61906]: DEBUG nova.network.neutron [req-435b8738-8e3e-4cbc-bb21-e19b3beee001 req-5aa4a4b2-92cf-4885-973d-aeab01f356d1 service nova] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Updating instance_info_cache with network_info: [{"id": "ef2c8975-d452-47be-8622-b0a062d3b4b1", "address": "fa:16:3e:f6:7c:45", "network": {"id": "aee89e54-f939-4aa4-9212-df451d432c3a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-702450817-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7849d220ba9f4c37a7c0c73535c17301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec1528b-3e87-477b-8ab2-02696ad47e66", "external-id": "nsx-vlan-transportzone-180", "segmentation_id": 180, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef2c8975-d4", "ovs_interfaceid": "ef2c8975-d452-47be-8622-b0a062d3b4b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.625691] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "refresh_cache-fdd638bd-00b0-40e7-92df-ba0a378a0ae6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.625870] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "refresh_cache-fdd638bd-00b0-40e7-92df-ba0a378a0ae6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.626015] env[61906]: DEBUG nova.network.neutron [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1086.935521] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357118, 'name': CreateVM_Task, 'duration_secs': 0.618808} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.936022] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1086.936564] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.936771] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.936923] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1086.937204] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c175d64-0bcc-436e-bbba-39117e8c7055 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.941921] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Waiting for the task: (returnval){ [ 1086.941921] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5271aaaa-1176-c8b1-20aa-1d2ec1b26d72" [ 1086.941921] env[61906]: _type = "Task" [ 1086.941921] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.951187] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5271aaaa-1176-c8b1-20aa-1d2ec1b26d72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.106976] env[61906]: DEBUG oslo_concurrency.lockutils [req-435b8738-8e3e-4cbc-bb21-e19b3beee001 req-5aa4a4b2-92cf-4885-973d-aeab01f356d1 service nova] Releasing lock "refresh_cache-1dae8ad9-245f-4b05-9c65-59c2e19af22b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.156909] env[61906]: DEBUG nova.network.neutron [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1087.210738] env[61906]: DEBUG nova.network.neutron [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance_info_cache with network_info: [{"id": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "address": "fa:16:3e:c2:38:09", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda2e07e3-74", "ovs_interfaceid": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.286624] env[61906]: DEBUG nova.network.neutron [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Updating instance_info_cache with network_info: [{"id": "fb600c17-7915-4f83-982e-605f7b6f1f61", "address": "fa:16:3e:e0:ee:b8", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb600c17-79", "ovs_interfaceid": "fb600c17-7915-4f83-982e-605f7b6f1f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.452168] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5271aaaa-1176-c8b1-20aa-1d2ec1b26d72, 'name': SearchDatastore_Task, 'duration_secs': 0.012207} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.452495] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.452736] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1087.452973] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.453140] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.453327] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.453594] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-537fb32e-2130-4da4-8ee4-7e9c40510966 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.461358] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.461546] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1087.462309] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a0edb82-0413-4f29-85ee-a77cdd6e581d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.466998] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Waiting for the task: (returnval){ [ 1087.466998] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f9ec5e-d39b-de63-8d65-def8aaec4c86" [ 1087.466998] env[61906]: _type = "Task" [ 1087.466998] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.474457] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f9ec5e-d39b-de63-8d65-def8aaec4c86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.713361] env[61906]: DEBUG oslo_concurrency.lockutils [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.789797] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "refresh_cache-fdd638bd-00b0-40e7-92df-ba0a378a0ae6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.790122] env[61906]: DEBUG nova.compute.manager [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Instance network_info: |[{"id": "fb600c17-7915-4f83-982e-605f7b6f1f61", "address": "fa:16:3e:e0:ee:b8", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb600c17-79", "ovs_interfaceid": "fb600c17-7915-4f83-982e-605f7b6f1f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1087.790572] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:ee:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1895250-76cc-41f7-b7f8-2e5679494607', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fb600c17-7915-4f83-982e-605f7b6f1f61', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1087.797986] env[61906]: DEBUG oslo.service.loopingcall [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1087.798530] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1087.798775] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-79b02a45-d903-43e0-9422-177342222756 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.818411] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1087.818411] env[61906]: value = "task-1357119" [ 1087.818411] env[61906]: _type = "Task" [ 1087.818411] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.825711] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357119, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.923625] env[61906]: DEBUG nova.compute.manager [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Received event network-vif-plugged-fb600c17-7915-4f83-982e-605f7b6f1f61 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.924095] env[61906]: DEBUG oslo_concurrency.lockutils [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] Acquiring lock "fdd638bd-00b0-40e7-92df-ba0a378a0ae6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.924095] env[61906]: DEBUG oslo_concurrency.lockutils [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] Lock "fdd638bd-00b0-40e7-92df-ba0a378a0ae6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.924270] env[61906]: DEBUG oslo_concurrency.lockutils [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] Lock "fdd638bd-00b0-40e7-92df-ba0a378a0ae6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.924445] env[61906]: DEBUG nova.compute.manager [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] No waiting events found dispatching network-vif-plugged-fb600c17-7915-4f83-982e-605f7b6f1f61 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1087.924623] env[61906]: WARNING nova.compute.manager [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Received unexpected event network-vif-plugged-fb600c17-7915-4f83-982e-605f7b6f1f61 for instance with vm_state building and task_state spawning. [ 1087.924785] env[61906]: DEBUG nova.compute.manager [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Received event network-changed-fb600c17-7915-4f83-982e-605f7b6f1f61 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.924946] env[61906]: DEBUG nova.compute.manager [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Refreshing instance network info cache due to event network-changed-fb600c17-7915-4f83-982e-605f7b6f1f61. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1087.925148] env[61906]: DEBUG oslo_concurrency.lockutils [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] Acquiring lock "refresh_cache-fdd638bd-00b0-40e7-92df-ba0a378a0ae6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.925293] env[61906]: DEBUG oslo_concurrency.lockutils [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] Acquired lock "refresh_cache-fdd638bd-00b0-40e7-92df-ba0a378a0ae6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.925462] env[61906]: DEBUG nova.network.neutron [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Refreshing network info cache for port fb600c17-7915-4f83-982e-605f7b6f1f61 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1087.977270] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f9ec5e-d39b-de63-8d65-def8aaec4c86, 'name': SearchDatastore_Task, 'duration_secs': 0.015595} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.978155] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ba94fca-696c-41a1-8f45-d8dab69c4307 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.982943] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Waiting for the task: (returnval){ [ 1087.982943] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521a62a7-fbef-324c-8133-4f6fd1ffe198" [ 1087.982943] env[61906]: _type = "Task" [ 1087.982943] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.990214] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521a62a7-fbef-324c-8133-4f6fd1ffe198, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.329811] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357119, 'name': CreateVM_Task, 'duration_secs': 0.293292} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.330058] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1088.330773] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.331007] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.331403] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1088.331694] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-340526b1-c512-4970-8e7e-7e4ebd3d7463 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.336206] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1088.336206] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c75335-a877-f755-a2aa-9ed4d27b3ca2" [ 1088.336206] env[61906]: _type = "Task" [ 1088.336206] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.343909] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c75335-a877-f755-a2aa-9ed4d27b3ca2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.496278] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]521a62a7-fbef-324c-8133-4f6fd1ffe198, 'name': SearchDatastore_Task, 'duration_secs': 0.008686} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.496546] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.496814] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 1dae8ad9-245f-4b05-9c65-59c2e19af22b/1dae8ad9-245f-4b05-9c65-59c2e19af22b.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1088.497089] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b51eb72-8075-46cc-9efb-cf9ca0e9f71f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.503985] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Waiting for the task: (returnval){ [ 1088.503985] env[61906]: value = "task-1357120" [ 1088.503985] env[61906]: _type = "Task" [ 1088.503985] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.513095] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357120, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.629262] env[61906]: DEBUG nova.network.neutron [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Updated VIF entry in instance network info cache for port fb600c17-7915-4f83-982e-605f7b6f1f61. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1088.629718] env[61906]: DEBUG nova.network.neutron [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Updating instance_info_cache with network_info: [{"id": "fb600c17-7915-4f83-982e-605f7b6f1f61", "address": "fa:16:3e:e0:ee:b8", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb600c17-79", "ovs_interfaceid": "fb600c17-7915-4f83-982e-605f7b6f1f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.847048] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c75335-a877-f755-a2aa-9ed4d27b3ca2, 'name': SearchDatastore_Task, 'duration_secs': 0.008599} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.847350] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.847654] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1088.847921] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.848091] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.848281] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1088.848575] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-702d1b6f-6c89-4890-bc5b-7c0b3853e668 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.862112] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1088.862344] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1088.863536] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfa85bdc-9e04-4a28-af70-702a95330f13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.869667] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1088.869667] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52030168-d17b-c007-9c84-fdb296ac7f4c" [ 1088.869667] env[61906]: _type = "Task" [ 1088.869667] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.878359] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52030168-d17b-c007-9c84-fdb296ac7f4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.013572] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357120, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460922} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.013905] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 1dae8ad9-245f-4b05-9c65-59c2e19af22b/1dae8ad9-245f-4b05-9c65-59c2e19af22b.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1089.014064] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1089.014318] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-286076ee-b6aa-4fe1-8a56-ea752d3677d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.020458] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Waiting for the task: (returnval){ [ 1089.020458] env[61906]: value = "task-1357121" [ 1089.020458] env[61906]: _type = "Task" [ 1089.020458] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.027878] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357121, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.132866] env[61906]: DEBUG oslo_concurrency.lockutils [req-0722ed7d-9fd2-4dfb-bf13-25d8f2f97b62 req-7ec986c8-177b-47b5-8b5b-a1aa5f39017b service nova] Releasing lock "refresh_cache-fdd638bd-00b0-40e7-92df-ba0a378a0ae6" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.229304] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37ff9cd-d091-4f77-9f49-44caee63ebe0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.247573] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance 'fc0e81b8-86b8-4396-afdb-fe9184815fc0' progress to 0 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1089.380359] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52030168-d17b-c007-9c84-fdb296ac7f4c, 'name': SearchDatastore_Task, 'duration_secs': 0.050221} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.381106] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7b865c6-96aa-4e6c-860f-5159c5b168f2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.385786] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1089.385786] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522a78b8-b9e2-fe41-89e3-313c50c5e8cf" [ 1089.385786] env[61906]: _type = "Task" [ 1089.385786] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.392985] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522a78b8-b9e2-fe41-89e3-313c50c5e8cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.530143] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357121, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077984} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.530430] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1089.531222] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a5d9b9-53b0-4c50-8dc4-39f691499e7a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.554622] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 1dae8ad9-245f-4b05-9c65-59c2e19af22b/1dae8ad9-245f-4b05-9c65-59c2e19af22b.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.554862] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4673a112-e66e-4ecb-aa2d-09a65874a26f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.573370] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Waiting for the task: (returnval){ [ 1089.573370] env[61906]: value = "task-1357122" [ 1089.573370] env[61906]: _type = "Task" [ 1089.573370] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.580606] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357122, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.753929] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1089.754267] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec84e908-63b4-4064-9cae-8fce7ce06528 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.761540] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1089.761540] env[61906]: value = "task-1357123" [ 1089.761540] env[61906]: _type = "Task" [ 1089.761540] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.770966] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357123, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.897067] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522a78b8-b9e2-fe41-89e3-313c50c5e8cf, 'name': SearchDatastore_Task, 'duration_secs': 0.008855} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.897067] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.897067] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] fdd638bd-00b0-40e7-92df-ba0a378a0ae6/fdd638bd-00b0-40e7-92df-ba0a378a0ae6.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1089.897556] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9fe709b7-8060-4e38-81a5-610d8eaa4c93 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.903817] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1089.903817] env[61906]: value = "task-1357124" [ 1089.903817] env[61906]: _type = "Task" [ 1089.903817] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.912411] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357124, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.084301] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357122, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.272355] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357123, 'name': PowerOffVM_Task, 'duration_secs': 0.193914} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.272672] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1090.272907] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance 'fc0e81b8-86b8-4396-afdb-fe9184815fc0' progress to 17 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1090.413764] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357124, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.584058] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357122, 'name': ReconfigVM_Task, 'duration_secs': 0.606036} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.584318] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 1dae8ad9-245f-4b05-9c65-59c2e19af22b/1dae8ad9-245f-4b05-9c65-59c2e19af22b.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1090.584939] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bdeb4569-26ff-4b4f-87fe-274ec338c5aa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.590918] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Waiting for the task: (returnval){ [ 1090.590918] env[61906]: value = "task-1357125" [ 1090.590918] env[61906]: _type = "Task" [ 1090.590918] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.599614] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357125, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.780472] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1090.780741] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1090.780892] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1090.781102] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1090.781283] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1090.781440] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1090.781656] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1090.781826] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1090.782014] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1090.782195] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1090.782380] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1090.787391] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4649db85-e528-4f7b-8990-981dddd9787c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.803552] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1090.803552] env[61906]: value = "task-1357126" [ 1090.803552] env[61906]: _type = "Task" [ 1090.803552] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.813522] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357126, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.915196] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357124, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.100641] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357125, 'name': Rename_Task, 'duration_secs': 0.322518} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.100957] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1091.101273] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-043b3df6-6843-4008-8d6d-0a699e64b0c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.106987] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Waiting for the task: (returnval){ [ 1091.106987] env[61906]: value = "task-1357127" [ 1091.106987] env[61906]: _type = "Task" [ 1091.106987] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.115526] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357127, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.312995] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357126, 'name': ReconfigVM_Task, 'duration_secs': 0.185813} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.313354] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance 'fc0e81b8-86b8-4396-afdb-fe9184815fc0' progress to 33 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1091.415548] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357124, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.617675] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357127, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.820427] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1091.820736] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1091.820950] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1091.821072] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1091.821225] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1091.821381] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1091.821591] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1091.821757] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1091.821959] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1091.822183] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1091.822385] env[61906]: DEBUG nova.virt.hardware [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1091.828231] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1091.828561] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba645646-5c8a-4c10-aa46-f5a5a8921b7b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.847385] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1091.847385] env[61906]: value = "task-1357128" [ 1091.847385] env[61906]: _type = "Task" [ 1091.847385] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.855528] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357128, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.916258] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357124, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.600767} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.916649] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] fdd638bd-00b0-40e7-92df-ba0a378a0ae6/fdd638bd-00b0-40e7-92df-ba0a378a0ae6.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1091.916944] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1091.917282] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e14e3e79-3fb3-4d8c-be36-485365928e0f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.924154] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1091.924154] env[61906]: value = "task-1357129" [ 1091.924154] env[61906]: _type = "Task" [ 1091.924154] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.932847] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357129, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.117767] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357127, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.357456] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357128, 'name': ReconfigVM_Task, 'duration_secs': 0.241544} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.357622] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1092.358373] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecfa14fa-43e7-4fad-9fe2-a993a6b785cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.379902] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] fc0e81b8-86b8-4396-afdb-fe9184815fc0/fc0e81b8-86b8-4396-afdb-fe9184815fc0.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1092.380467] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-941ad85a-71dc-4037-8000-0b4c74393210 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.403846] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1092.403846] env[61906]: value = "task-1357130" [ 1092.403846] env[61906]: _type = "Task" [ 1092.403846] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.411346] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357130, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.432102] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357129, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073818} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.432362] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1092.433163] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44009adf-8263-4a1d-864c-c2c6e473aa85 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.457635] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] fdd638bd-00b0-40e7-92df-ba0a378a0ae6/fdd638bd-00b0-40e7-92df-ba0a378a0ae6.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1092.458028] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8be5ca0a-dfef-4b18-bd1c-3feced3a8ee2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.478234] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1092.478234] env[61906]: value = "task-1357131" [ 1092.478234] env[61906]: _type = "Task" [ 1092.478234] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.486587] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357131, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.622058] env[61906]: DEBUG oslo_vmware.api [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357127, 'name': PowerOnVM_Task, 'duration_secs': 1.064547} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.622058] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1092.622306] env[61906]: INFO nova.compute.manager [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Took 9.02 seconds to spawn the instance on the hypervisor. [ 1092.622506] env[61906]: DEBUG nova.compute.manager [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1092.623588] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ee5244-cb02-4a08-84c4-1b941d6eaba7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.913645] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357130, 'name': ReconfigVM_Task, 'duration_secs': 0.29807} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.913956] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Reconfigured VM instance instance-00000065 to attach disk [datastore2] fc0e81b8-86b8-4396-afdb-fe9184815fc0/fc0e81b8-86b8-4396-afdb-fe9184815fc0.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1092.914250] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance 'fc0e81b8-86b8-4396-afdb-fe9184815fc0' progress to 50 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1092.987962] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357131, 'name': ReconfigVM_Task, 'duration_secs': 0.306373} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.988285] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Reconfigured VM instance instance-00000069 to attach disk [datastore2] fdd638bd-00b0-40e7-92df-ba0a378a0ae6/fdd638bd-00b0-40e7-92df-ba0a378a0ae6.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1092.988989] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-28ac68ec-3dc5-4f40-893d-4d45461fa6e5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.994983] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1092.994983] env[61906]: value = "task-1357132" [ 1092.994983] env[61906]: _type = "Task" [ 1092.994983] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.002365] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357132, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.142325] env[61906]: INFO nova.compute.manager [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Took 13.77 seconds to build instance. [ 1093.420609] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432cc0d6-35e1-401b-b7ac-66d48a17f03a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.440428] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3fbcaa4-973c-4602-be29-a47ec600d855 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.458634] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance 'fc0e81b8-86b8-4396-afdb-fe9184815fc0' progress to 67 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1093.504904] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357132, 'name': Rename_Task, 'duration_secs': 0.171212} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.505197] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1093.505440] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a29e54a9-c76b-4a1f-b6de-c2e342b98c91 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.511402] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1093.511402] env[61906]: value = "task-1357133" [ 1093.511402] env[61906]: _type = "Task" [ 1093.511402] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.518538] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357133, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.644955] env[61906]: DEBUG oslo_concurrency.lockutils [None req-aad9657c-659e-433c-9664-14d06b2aae80 tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.282s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.781313] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquiring lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.781572] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.781820] env[61906]: INFO nova.compute.manager [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Rebooting instance [ 1093.996250] env[61906]: DEBUG nova.network.neutron [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Port da2e07e3-7453-4d91-88f7-26247bd7ea9d binding to destination host cpu-1 is already ACTIVE {{(pid=61906) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1094.022342] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357133, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.370470] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquiring lock "refresh_cache-1dae8ad9-245f-4b05-9c65-59c2e19af22b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.370761] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquired lock "refresh_cache-1dae8ad9-245f-4b05-9c65-59c2e19af22b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.370871] env[61906]: DEBUG nova.network.neutron [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1094.521601] env[61906]: DEBUG oslo_vmware.api [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357133, 'name': PowerOnVM_Task, 'duration_secs': 0.610388} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.521884] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1094.522108] env[61906]: INFO nova.compute.manager [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Took 8.69 seconds to spawn the instance on the hypervisor. [ 1094.522299] env[61906]: DEBUG nova.compute.manager [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1094.523054] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081ab9fc-c73a-41f1-9da5-47a86f4d2c81 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.019869] env[61906]: DEBUG oslo_concurrency.lockutils [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.020128] env[61906]: DEBUG oslo_concurrency.lockutils [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.020335] env[61906]: DEBUG oslo_concurrency.lockutils [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.040505] env[61906]: INFO nova.compute.manager [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Took 15.32 seconds to build instance. [ 1095.088687] env[61906]: DEBUG nova.network.neutron [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Updating instance_info_cache with network_info: [{"id": "ef2c8975-d452-47be-8622-b0a062d3b4b1", "address": "fa:16:3e:f6:7c:45", "network": {"id": "aee89e54-f939-4aa4-9212-df451d432c3a", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-702450817-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7849d220ba9f4c37a7c0c73535c17301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec1528b-3e87-477b-8ab2-02696ad47e66", "external-id": "nsx-vlan-transportzone-180", "segmentation_id": 180, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef2c8975-d4", "ovs_interfaceid": "ef2c8975-d452-47be-8622-b0a062d3b4b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.542736] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b8bef1a5-e1f3-4e21-b27f-27cf7112eee9 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "fdd638bd-00b0-40e7-92df-ba0a378a0ae6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.827s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.592929] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Releasing lock "refresh_cache-1dae8ad9-245f-4b05-9c65-59c2e19af22b" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.594458] env[61906]: DEBUG nova.compute.manager [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1095.595336] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f225c7-8e08-4844-8a7b-ecf18ab4a912 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.080789] env[61906]: DEBUG oslo_concurrency.lockutils [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.080995] env[61906]: DEBUG oslo_concurrency.lockutils [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.081193] env[61906]: DEBUG nova.network.neutron [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1096.351063] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "fdd638bd-00b0-40e7-92df-ba0a378a0ae6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.351063] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "fdd638bd-00b0-40e7-92df-ba0a378a0ae6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.351063] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "fdd638bd-00b0-40e7-92df-ba0a378a0ae6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.351063] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "fdd638bd-00b0-40e7-92df-ba0a378a0ae6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.351298] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "fdd638bd-00b0-40e7-92df-ba0a378a0ae6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.353305] env[61906]: INFO nova.compute.manager [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Terminating instance [ 1096.354997] env[61906]: DEBUG nova.compute.manager [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1096.355215] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1096.356052] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49ba1c2-ee40-487b-b8e7-8f50c3065f2b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.363271] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1096.363504] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d2269173-c521-42fd-9862-88d721099af7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.369591] env[61906]: DEBUG oslo_vmware.api [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1096.369591] env[61906]: value = "task-1357134" [ 1096.369591] env[61906]: _type = "Task" [ 1096.369591] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.377172] env[61906]: DEBUG oslo_vmware.api [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357134, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.612576] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d272fd2-7d1e-44cc-aecc-ea0e004af1b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.622112] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Doing hard reboot of VM {{(pid=61906) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 1096.622223] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-7583cef2-098d-46f8-9c29-e65621ce6a1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.628140] env[61906]: DEBUG oslo_vmware.api [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Waiting for the task: (returnval){ [ 1096.628140] env[61906]: value = "task-1357135" [ 1096.628140] env[61906]: _type = "Task" [ 1096.628140] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.635419] env[61906]: DEBUG oslo_vmware.api [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357135, 'name': ResetVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.798363] env[61906]: DEBUG nova.network.neutron [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance_info_cache with network_info: [{"id": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "address": "fa:16:3e:c2:38:09", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda2e07e3-74", "ovs_interfaceid": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.881049] env[61906]: DEBUG oslo_vmware.api [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357134, 'name': PowerOffVM_Task, 'duration_secs': 0.193604} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.881367] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1096.881550] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1096.881828] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-db21982e-c5b9-4a09-9e0c-ab945bb70e06 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.941129] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1096.941384] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1096.941564] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleting the datastore file [datastore2] fdd638bd-00b0-40e7-92df-ba0a378a0ae6 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1096.941828] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a811b1e-d477-4838-b70c-10e03a929e92 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.948156] env[61906]: DEBUG oslo_vmware.api [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1096.948156] env[61906]: value = "task-1357137" [ 1096.948156] env[61906]: _type = "Task" [ 1096.948156] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.957162] env[61906]: DEBUG oslo_vmware.api [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357137, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.137744] env[61906]: DEBUG oslo_vmware.api [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357135, 'name': ResetVM_Task, 'duration_secs': 0.08519} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.137983] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Did hard reboot of VM {{(pid=61906) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 1097.138194] env[61906]: DEBUG nova.compute.manager [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1097.138931] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b86a3f-3660-428e-81e8-1513036d4f29 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.301586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.457545] env[61906]: DEBUG oslo_vmware.api [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357137, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149842} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.457817] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1097.458015] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1097.458209] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1097.458394] env[61906]: INFO nova.compute.manager [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1097.458642] env[61906]: DEBUG oslo.service.loopingcall [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1097.458841] env[61906]: DEBUG nova.compute.manager [-] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1097.458932] env[61906]: DEBUG nova.network.neutron [-] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1097.650050] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4dfe2d27-6470-4217-b992-cb25c5d6ad3e tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.868s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.720385] env[61906]: DEBUG nova.compute.manager [req-7da7a469-231f-4d01-abe2-11676726c6b5 req-e41921cd-82c1-4dbe-b300-e475ecacf6ca service nova] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Received event network-vif-deleted-fb600c17-7915-4f83-982e-605f7b6f1f61 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1097.720385] env[61906]: INFO nova.compute.manager [req-7da7a469-231f-4d01-abe2-11676726c6b5 req-e41921cd-82c1-4dbe-b300-e475ecacf6ca service nova] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Neutron deleted interface fb600c17-7915-4f83-982e-605f7b6f1f61; detaching it from the instance and deleting it from the info cache [ 1097.720487] env[61906]: DEBUG nova.network.neutron [req-7da7a469-231f-4d01-abe2-11676726c6b5 req-e41921cd-82c1-4dbe-b300-e475ecacf6ca service nova] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.831267] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d2bc3dd-b737-487f-afb2-77b496731cf3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.851153] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f20edd4-e580-46d5-8704-61449216084d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.858915] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance 'fc0e81b8-86b8-4396-afdb-fe9184815fc0' progress to 83 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1098.199974] env[61906]: DEBUG nova.network.neutron [-] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.223470] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8a39653b-b84c-48d0-a4dd-3584bbf32ff4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.233668] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c74af1-4fc8-4a0d-a35f-c2269aecbd1d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.260171] env[61906]: DEBUG nova.compute.manager [req-7da7a469-231f-4d01-abe2-11676726c6b5 req-e41921cd-82c1-4dbe-b300-e475ecacf6ca service nova] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Detach interface failed, port_id=fb600c17-7915-4f83-982e-605f7b6f1f61, reason: Instance fdd638bd-00b0-40e7-92df-ba0a378a0ae6 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1098.365539] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1098.365860] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10e0bb31-7278-4a59-b787-1c5375487835 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.372808] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1098.372808] env[61906]: value = "task-1357138" [ 1098.372808] env[61906]: _type = "Task" [ 1098.372808] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.380536] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357138, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.406563] env[61906]: DEBUG oslo_concurrency.lockutils [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquiring lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.406749] env[61906]: DEBUG oslo_concurrency.lockutils [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.407041] env[61906]: DEBUG oslo_concurrency.lockutils [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquiring lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.407312] env[61906]: DEBUG oslo_concurrency.lockutils [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.407557] env[61906]: DEBUG oslo_concurrency.lockutils [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.409849] env[61906]: INFO nova.compute.manager [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Terminating instance [ 1098.411812] env[61906]: DEBUG nova.compute.manager [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1098.412076] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1098.413016] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a15483b7-71ea-48e1-976f-61c55ada3afa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.421492] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1098.421743] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a78cd37-74f1-481f-bbd3-02d064aae52e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.427188] env[61906]: DEBUG oslo_vmware.api [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Waiting for the task: (returnval){ [ 1098.427188] env[61906]: value = "task-1357139" [ 1098.427188] env[61906]: _type = "Task" [ 1098.427188] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.435296] env[61906]: DEBUG oslo_vmware.api [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357139, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.702936] env[61906]: INFO nova.compute.manager [-] [instance: fdd638bd-00b0-40e7-92df-ba0a378a0ae6] Took 1.24 seconds to deallocate network for instance. [ 1098.882059] env[61906]: DEBUG oslo_vmware.api [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357138, 'name': PowerOnVM_Task, 'duration_secs': 0.369431} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.882349] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1098.882556] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-34f83f2e-78e6-409d-8d26-8266fbfe5d6f tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance 'fc0e81b8-86b8-4396-afdb-fe9184815fc0' progress to 100 {{(pid=61906) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1098.937342] env[61906]: DEBUG oslo_vmware.api [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357139, 'name': PowerOffVM_Task, 'duration_secs': 0.198809} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.937649] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1098.937832] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1098.938113] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6c0196db-ced9-4b90-b29d-d00d06091a57 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.002649] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1099.002902] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1099.003109] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Deleting the datastore file [datastore2] 1dae8ad9-245f-4b05-9c65-59c2e19af22b {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1099.003384] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-539266f8-0528-4e06-a959-00172467cb69 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.009482] env[61906]: DEBUG oslo_vmware.api [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Waiting for the task: (returnval){ [ 1099.009482] env[61906]: value = "task-1357141" [ 1099.009482] env[61906]: _type = "Task" [ 1099.009482] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.016755] env[61906]: DEBUG oslo_vmware.api [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357141, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.211247] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.211632] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.211950] env[61906]: DEBUG nova.objects.instance [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lazy-loading 'resources' on Instance uuid fdd638bd-00b0-40e7-92df-ba0a378a0ae6 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.519612] env[61906]: DEBUG oslo_vmware.api [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Task: {'id': task-1357141, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205978} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.519896] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1099.520105] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1099.520290] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1099.520468] env[61906]: INFO nova.compute.manager [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1099.520708] env[61906]: DEBUG oslo.service.loopingcall [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1099.520908] env[61906]: DEBUG nova.compute.manager [-] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1099.521009] env[61906]: DEBUG nova.network.neutron [-] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1099.781018] env[61906]: DEBUG nova.compute.manager [req-8d27bf12-cbee-44f9-8568-367ada94a145 req-be244497-5af5-4b2b-ad81-fcec817d6afb service nova] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Received event network-vif-deleted-ef2c8975-d452-47be-8622-b0a062d3b4b1 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1099.781262] env[61906]: INFO nova.compute.manager [req-8d27bf12-cbee-44f9-8568-367ada94a145 req-be244497-5af5-4b2b-ad81-fcec817d6afb service nova] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Neutron deleted interface ef2c8975-d452-47be-8622-b0a062d3b4b1; detaching it from the instance and deleting it from the info cache [ 1099.781404] env[61906]: DEBUG nova.network.neutron [req-8d27bf12-cbee-44f9-8568-367ada94a145 req-be244497-5af5-4b2b-ad81-fcec817d6afb service nova] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.816484] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c60a9860-6b98-4235-bc26-e50cd7c45bec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.824420] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5027830b-1f96-4b6c-ab00-3a3926b9df0b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.865805] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d402d61-2552-41fa-a832-6f5f9718774e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.873827] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea0c8b6-1441-4792-96f2-8c855592c104 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.886989] env[61906]: DEBUG nova.compute.provider_tree [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.261036] env[61906]: DEBUG nova.network.neutron [-] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.284197] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5b7707a-ce4f-4a35-adcd-b86163b9d53e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.294170] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f87cc330-03ca-42e4-9012-6ff71a10ead1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.320797] env[61906]: DEBUG nova.compute.manager [req-8d27bf12-cbee-44f9-8568-367ada94a145 req-be244497-5af5-4b2b-ad81-fcec817d6afb service nova] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Detach interface failed, port_id=ef2c8975-d452-47be-8622-b0a062d3b4b1, reason: Instance 1dae8ad9-245f-4b05-9c65-59c2e19af22b could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1100.389790] env[61906]: DEBUG nova.scheduler.client.report [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1100.763487] env[61906]: INFO nova.compute.manager [-] [instance: 1dae8ad9-245f-4b05-9c65-59c2e19af22b] Took 1.24 seconds to deallocate network for instance. [ 1100.895054] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.683s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.915168] env[61906]: INFO nova.scheduler.client.report [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted allocations for instance fdd638bd-00b0-40e7-92df-ba0a378a0ae6 [ 1101.270482] env[61906]: DEBUG oslo_concurrency.lockutils [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.270775] env[61906]: DEBUG oslo_concurrency.lockutils [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.271108] env[61906]: DEBUG nova.objects.instance [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Lazy-loading 'resources' on Instance uuid 1dae8ad9-245f-4b05-9c65-59c2e19af22b {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.423805] env[61906]: DEBUG oslo_concurrency.lockutils [None req-b63a9024-ec1f-4897-8542-fc89a6046c1d tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "fdd638bd-00b0-40e7-92df-ba0a378a0ae6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.073s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.653978] env[61906]: DEBUG nova.network.neutron [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Port da2e07e3-7453-4d91-88f7-26247bd7ea9d binding to destination host cpu-1 is already ACTIVE {{(pid=61906) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1101.654386] env[61906]: DEBUG oslo_concurrency.lockutils [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.654796] env[61906]: DEBUG oslo_concurrency.lockutils [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.655090] env[61906]: DEBUG nova.network.neutron [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1101.819030] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "f235d418-6dbb-41ac-8a7b-d959f31adc22" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.819293] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "f235d418-6dbb-41ac-8a7b-d959f31adc22" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.819557] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "f235d418-6dbb-41ac-8a7b-d959f31adc22-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.819764] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "f235d418-6dbb-41ac-8a7b-d959f31adc22-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.819943] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "f235d418-6dbb-41ac-8a7b-d959f31adc22-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.822121] env[61906]: INFO nova.compute.manager [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Terminating instance [ 1101.823909] env[61906]: DEBUG nova.compute.manager [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1101.824122] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1101.824954] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7851353b-c842-43bf-b87b-e328fd148444 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.832956] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1101.835821] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01af438a-19ec-409f-9840-b88ede856ecd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.843038] env[61906]: DEBUG oslo_vmware.api [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1101.843038] env[61906]: value = "task-1357142" [ 1101.843038] env[61906]: _type = "Task" [ 1101.843038] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.850621] env[61906]: DEBUG oslo_vmware.api [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357142, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.880325] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e35753-477a-44cb-b4c1-c9182963ad7b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.887845] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1763faf6-bf77-4b06-b841-ce3ab59744bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.919221] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739e9135-633e-402a-8a10-ea08ff0f87d4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.927876] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd663099-e99d-403b-b895-fcb573c3221f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.942704] env[61906]: DEBUG nova.compute.provider_tree [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1102.353879] env[61906]: DEBUG oslo_vmware.api [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357142, 'name': PowerOffVM_Task, 'duration_secs': 0.464345} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.354261] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1102.354455] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1102.354717] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dcc2ceae-7980-48ef-943a-f9e1c7dee6ed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.386502] env[61906]: DEBUG nova.network.neutron [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance_info_cache with network_info: [{"id": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "address": "fa:16:3e:c2:38:09", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda2e07e3-74", "ovs_interfaceid": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.431524] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1102.431642] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1102.431750] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleting the datastore file [datastore1] f235d418-6dbb-41ac-8a7b-d959f31adc22 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1102.432070] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-61306936-f7ee-420c-a97b-2b830758734c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.440634] env[61906]: DEBUG oslo_vmware.api [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1102.440634] env[61906]: value = "task-1357144" [ 1102.440634] env[61906]: _type = "Task" [ 1102.440634] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.445508] env[61906]: DEBUG nova.scheduler.client.report [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1102.451461] env[61906]: DEBUG oslo_vmware.api [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357144, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.889184] env[61906]: DEBUG oslo_concurrency.lockutils [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.949825] env[61906]: DEBUG oslo_vmware.api [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357144, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17133} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.950218] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1102.950351] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1102.950556] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1102.950738] env[61906]: INFO nova.compute.manager [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1102.951047] env[61906]: DEBUG oslo.service.loopingcall [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1102.951264] env[61906]: DEBUG nova.compute.manager [-] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1102.951358] env[61906]: DEBUG nova.network.neutron [-] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1102.953401] env[61906]: DEBUG oslo_concurrency.lockutils [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.683s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.973499] env[61906]: INFO nova.scheduler.client.report [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Deleted allocations for instance 1dae8ad9-245f-4b05-9c65-59c2e19af22b [ 1103.208971] env[61906]: DEBUG nova.compute.manager [req-e1a64bd9-8844-4c89-acdc-7e982cce29df req-d72997b3-9a43-401d-9a38-97ddc2580637 service nova] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Received event network-vif-deleted-213895ba-b477-4936-a766-369ea1f71948 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1103.209057] env[61906]: INFO nova.compute.manager [req-e1a64bd9-8844-4c89-acdc-7e982cce29df req-d72997b3-9a43-401d-9a38-97ddc2580637 service nova] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Neutron deleted interface 213895ba-b477-4936-a766-369ea1f71948; detaching it from the instance and deleting it from the info cache [ 1103.209252] env[61906]: DEBUG nova.network.neutron [req-e1a64bd9-8844-4c89-acdc-7e982cce29df req-d72997b3-9a43-401d-9a38-97ddc2580637 service nova] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.392279] env[61906]: DEBUG nova.compute.manager [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61906) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1103.392524] env[61906]: DEBUG oslo_concurrency.lockutils [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.392772] env[61906]: DEBUG oslo_concurrency.lockutils [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.481182] env[61906]: DEBUG oslo_concurrency.lockutils [None req-83545f21-3d7f-4120-b337-0075a1c63ffa tempest-InstanceActionsTestJSON-540313470 tempest-InstanceActionsTestJSON-540313470-project-member] Lock "1dae8ad9-245f-4b05-9c65-59c2e19af22b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.074s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.688403] env[61906]: DEBUG nova.network.neutron [-] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.712897] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04bc8d15-c4f2-4a3b-9bc8-5e10a9e13223 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.722446] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16af9c4-a4dd-41d2-899f-44a2af32487d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.746767] env[61906]: DEBUG nova.compute.manager [req-e1a64bd9-8844-4c89-acdc-7e982cce29df req-d72997b3-9a43-401d-9a38-97ddc2580637 service nova] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Detach interface failed, port_id=213895ba-b477-4936-a766-369ea1f71948, reason: Instance f235d418-6dbb-41ac-8a7b-d959f31adc22 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1103.895845] env[61906]: DEBUG nova.objects.instance [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lazy-loading 'migration_context' on Instance uuid fc0e81b8-86b8-4396-afdb-fe9184815fc0 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1104.191163] env[61906]: INFO nova.compute.manager [-] [instance: f235d418-6dbb-41ac-8a7b-d959f31adc22] Took 1.24 seconds to deallocate network for instance. [ 1104.475098] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650e5295-5021-496c-988e-c63b704babc7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.483896] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232c94a4-d0f4-43f9-9d96-ee0b24b2a9a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.516035] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f8de20-ab01-4505-9421-ed549a4927ba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.523525] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0437486f-873d-4faf-9de7-de79afd8c459 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.539114] env[61906]: DEBUG nova.compute.provider_tree [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.698431] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.045105] env[61906]: DEBUG nova.scheduler.client.report [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1106.057744] env[61906]: DEBUG oslo_concurrency.lockutils [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.664s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.068113] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.370s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.068539] env[61906]: DEBUG nova.objects.instance [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lazy-loading 'resources' on Instance uuid f235d418-6dbb-41ac-8a7b-d959f31adc22 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1106.680907] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8b17bc-f5fc-4916-a4bf-fe46ca45a2a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.690112] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f866c9f-6c7a-4739-bfca-05f497933d56 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.725559] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24fd3f2e-35ef-43cc-bb54-abbc7168f9e8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.733829] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549b5688-3080-4f14-9df7-7407ad06d862 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.749233] env[61906]: DEBUG nova.compute.provider_tree [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1107.256032] env[61906]: DEBUG nova.scheduler.client.report [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1107.611130] env[61906]: INFO nova.compute.manager [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Swapping old allocation on dict_keys(['6e524f5d-985e-41eb-a7ae-996be3ae1c1a']) held by migration 06224cea-26d8-443a-b299-09b795928f14 for instance [ 1107.637206] env[61906]: DEBUG nova.scheduler.client.report [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Overwriting current allocation {'allocations': {'6e524f5d-985e-41eb-a7ae-996be3ae1c1a': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 131}}, 'project_id': '0c6414fa082f49c2a568a772a256a8e5', 'user_id': '8ba1658b4ade413b98834102c39a255c', 'consumer_generation': 1} on consumer fc0e81b8-86b8-4396-afdb-fe9184815fc0 {{(pid=61906) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1107.720047] env[61906]: DEBUG oslo_concurrency.lockutils [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.720047] env[61906]: DEBUG oslo_concurrency.lockutils [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.720047] env[61906]: DEBUG nova.network.neutron [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1107.760170] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.692s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.777027] env[61906]: INFO nova.scheduler.client.report [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted allocations for instance f235d418-6dbb-41ac-8a7b-d959f31adc22 [ 1108.234902] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Acquiring lock "9f78bfe4-f8f2-44c5-a704-5e3763822211" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.235168] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Lock "9f78bfe4-f8f2-44c5-a704-5e3763822211" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.285104] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2837869e-3cb9-4d71-a3c6-8dd5b85a36bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "f235d418-6dbb-41ac-8a7b-d959f31adc22" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.466s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.457249] env[61906]: DEBUG nova.network.neutron [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance_info_cache with network_info: [{"id": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "address": "fa:16:3e:c2:38:09", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda2e07e3-74", "ovs_interfaceid": "da2e07e3-7453-4d91-88f7-26247bd7ea9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.737839] env[61906]: DEBUG nova.compute.manager [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1108.807836] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "43530cb7-293f-4b50-bf9b-e8ff1b047484" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.807987] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "43530cb7-293f-4b50-bf9b-e8ff1b047484" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.959963] env[61906]: DEBUG oslo_concurrency.lockutils [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-fc0e81b8-86b8-4396-afdb-fe9184815fc0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.960442] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1108.960769] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06a41df8-15fe-4cb4-9237-66600a73c9e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.967900] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1108.967900] env[61906]: value = "task-1357145" [ 1108.967900] env[61906]: _type = "Task" [ 1108.967900] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.976574] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357145, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.259749] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.260074] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.261929] env[61906]: INFO nova.compute.claims [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1109.310333] env[61906]: DEBUG nova.compute.manager [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1109.477896] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357145, 'name': PowerOffVM_Task, 'duration_secs': 0.196087} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.478257] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1109.478974] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1109.479230] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1109.479398] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1109.479620] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1109.479789] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1109.479943] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1109.480173] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1109.480340] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1109.480537] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1109.480680] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1109.480862] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1109.486231] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1059d5e-c280-4849-9ff6-bef148717a6c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.501608] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1109.501608] env[61906]: value = "task-1357146" [ 1109.501608] env[61906]: _type = "Task" [ 1109.501608] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.509721] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357146, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.830757] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.011903] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357146, 'name': ReconfigVM_Task, 'duration_secs': 0.126124} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.012896] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c6caac-6c0b-4ef5-9910-e8e094168e2b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.030635] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1110.030903] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1110.031089] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1110.031285] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1110.031437] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1110.031590] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1110.031801] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1110.031961] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1110.032164] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1110.032387] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1110.032574] env[61906]: DEBUG nova.virt.hardware [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1110.033339] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7264c2b-b00e-49c1-aae5-211e8db42154 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.038646] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1110.038646] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522e1e3f-1ac8-caff-f369-16a7966cf9bb" [ 1110.038646] env[61906]: _type = "Task" [ 1110.038646] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.046094] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522e1e3f-1ac8-caff-f369-16a7966cf9bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.340658] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267b8a90-cd82-4c51-a7af-437d744ca767 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.348102] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab09bea-1772-4659-bd1f-6202c8309b4c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.378341] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f7fc66-c820-41d9-8705-d24844a3c913 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.385075] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa5959f-4119-464a-a700-2099ffb9b309 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.397500] env[61906]: DEBUG nova.compute.provider_tree [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.548381] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]522e1e3f-1ac8-caff-f369-16a7966cf9bb, 'name': SearchDatastore_Task, 'duration_secs': 0.009405} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.553642] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1110.553918] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4bec0654-fa52-4ba1-af39-0d57e4b16e4b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.571615] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1110.571615] env[61906]: value = "task-1357147" [ 1110.571615] env[61906]: _type = "Task" [ 1110.571615] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.578998] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357147, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.900776] env[61906]: DEBUG nova.scheduler.client.report [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1110.923570] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.923851] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.924116] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.924383] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.924613] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.924798] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1110.924991] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1111.081382] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357147, 'name': ReconfigVM_Task, 'duration_secs': 0.17832} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.081665] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=61906) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1111.082454] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d54168b-7596-4fd6-8e45-39159d372642 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.104649] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] fc0e81b8-86b8-4396-afdb-fe9184815fc0/fc0e81b8-86b8-4396-afdb-fe9184815fc0.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1111.104925] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5f34bdf-53f7-4552-b804-1dcef1a3c888 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.122928] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1111.122928] env[61906]: value = "task-1357148" [ 1111.122928] env[61906]: _type = "Task" [ 1111.122928] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.130302] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357148, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.405896] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.146s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.406508] env[61906]: DEBUG nova.compute.manager [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1111.409173] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.579s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.410632] env[61906]: INFO nova.compute.claims [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1111.427887] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.632264] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357148, 'name': ReconfigVM_Task, 'duration_secs': 0.267963} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.632562] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Reconfigured VM instance instance-00000065 to attach disk [datastore2] fc0e81b8-86b8-4396-afdb-fe9184815fc0/fc0e81b8-86b8-4396-afdb-fe9184815fc0.vmdk or device None with type thin {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1111.633416] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa859083-c337-4891-8ec1-ce8d348b42d3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.650450] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a6c41e-face-4d59-8d70-f044a450d37c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.667625] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6ecfcc-46a3-4cf6-85f6-b82af931c035 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.684680] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea24af4-d8dd-4f33-bf71-173298d4d897 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.690863] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1111.691103] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4af0a38-0b16-4168-b1f6-0dce0dc22528 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.696783] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1111.696783] env[61906]: value = "task-1357149" [ 1111.696783] env[61906]: _type = "Task" [ 1111.696783] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.703694] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357149, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.915122] env[61906]: DEBUG nova.compute.utils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1111.918484] env[61906]: DEBUG nova.compute.manager [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1111.918652] env[61906]: DEBUG nova.network.neutron [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1111.962570] env[61906]: DEBUG nova.policy [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eeae160226364ec899385530c4bfefd1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7fef135be23341e9a47f5b22370013a5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1112.202588] env[61906]: DEBUG nova.network.neutron [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Successfully created port: 1e0e008f-84b1-42c4-a2d8-9663a0bb4b63 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1112.207833] env[61906]: DEBUG oslo_vmware.api [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357149, 'name': PowerOnVM_Task, 'duration_secs': 0.332465} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.208089] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1112.419478] env[61906]: DEBUG nova.compute.manager [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1112.511616] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f77a91c-5cd8-4e96-9b8d-c240e910cbbc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.520336] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fb388a6-4fcf-4bf9-a795-70d31400b276 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.548424] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fadc248e-eae1-4e11-a20e-407fd4ccd875 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.556191] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0050254-08f9-4893-846d-e04908491c09 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.568881] env[61906]: DEBUG nova.compute.provider_tree [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1113.071582] env[61906]: DEBUG nova.scheduler.client.report [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1113.220889] env[61906]: INFO nova.compute.manager [None req-308a91bf-2fb6-4d73-89d7-dc2502edb124 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance to original state: 'active' [ 1113.432718] env[61906]: DEBUG nova.compute.manager [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1113.460250] env[61906]: DEBUG nova.virt.hardware [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1113.460538] env[61906]: DEBUG nova.virt.hardware [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1113.460720] env[61906]: DEBUG nova.virt.hardware [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1113.460915] env[61906]: DEBUG nova.virt.hardware [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1113.461082] env[61906]: DEBUG nova.virt.hardware [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1113.461236] env[61906]: DEBUG nova.virt.hardware [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1113.461449] env[61906]: DEBUG nova.virt.hardware [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1113.461616] env[61906]: DEBUG nova.virt.hardware [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1113.461873] env[61906]: DEBUG nova.virt.hardware [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1113.461987] env[61906]: DEBUG nova.virt.hardware [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1113.462196] env[61906]: DEBUG nova.virt.hardware [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1113.463061] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b583dc-9a49-41e9-b78b-5d460dfca0cb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.470929] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60cf55cd-8d08-46b9-9766-1a43e3a8f4e3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.576891] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.168s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.577527] env[61906]: DEBUG nova.compute.manager [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1113.580120] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.152s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.580347] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.580535] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1113.581587] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba61da0b-1faf-4d83-9f6d-f6f5ceeba3ed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.591156] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f2dd950-91ad-487e-9b69-9fcb0c16a539 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.603815] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7eb71c-e5b9-4511-8799-724a8985f07c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.610936] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1f1168-e2a1-4895-8920-7e096da4e5fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.639584] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180580MB free_disk=150GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1113.639736] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.639891] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.742126] env[61906]: DEBUG nova.compute.manager [req-8f4a4c1f-2a8d-41c5-90e9-275cd97493c0 req-71b1ea48-0264-4722-92c5-b0a2b52ddbc3 service nova] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Received event network-vif-plugged-1e0e008f-84b1-42c4-a2d8-9663a0bb4b63 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1113.742126] env[61906]: DEBUG oslo_concurrency.lockutils [req-8f4a4c1f-2a8d-41c5-90e9-275cd97493c0 req-71b1ea48-0264-4722-92c5-b0a2b52ddbc3 service nova] Acquiring lock "9f78bfe4-f8f2-44c5-a704-5e3763822211-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.742126] env[61906]: DEBUG oslo_concurrency.lockutils [req-8f4a4c1f-2a8d-41c5-90e9-275cd97493c0 req-71b1ea48-0264-4722-92c5-b0a2b52ddbc3 service nova] Lock "9f78bfe4-f8f2-44c5-a704-5e3763822211-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.742126] env[61906]: DEBUG oslo_concurrency.lockutils [req-8f4a4c1f-2a8d-41c5-90e9-275cd97493c0 req-71b1ea48-0264-4722-92c5-b0a2b52ddbc3 service nova] Lock "9f78bfe4-f8f2-44c5-a704-5e3763822211-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.742126] env[61906]: DEBUG nova.compute.manager [req-8f4a4c1f-2a8d-41c5-90e9-275cd97493c0 req-71b1ea48-0264-4722-92c5-b0a2b52ddbc3 service nova] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] No waiting events found dispatching network-vif-plugged-1e0e008f-84b1-42c4-a2d8-9663a0bb4b63 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1113.742126] env[61906]: WARNING nova.compute.manager [req-8f4a4c1f-2a8d-41c5-90e9-275cd97493c0 req-71b1ea48-0264-4722-92c5-b0a2b52ddbc3 service nova] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Received unexpected event network-vif-plugged-1e0e008f-84b1-42c4-a2d8-9663a0bb4b63 for instance with vm_state building and task_state spawning. [ 1113.814407] env[61906]: DEBUG nova.network.neutron [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Successfully updated port: 1e0e008f-84b1-42c4-a2d8-9663a0bb4b63 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1114.082446] env[61906]: DEBUG nova.compute.utils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1114.084042] env[61906]: DEBUG nova.compute.manager [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1114.084604] env[61906]: DEBUG nova.network.neutron [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1114.140545] env[61906]: DEBUG nova.policy [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e28284ad5ce450597ef12c65c902e1c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ac4f2c9065c46068c7837eb85cd7e20', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1114.285360] env[61906]: DEBUG oslo_concurrency.lockutils [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.285629] env[61906]: DEBUG oslo_concurrency.lockutils [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.285862] env[61906]: DEBUG oslo_concurrency.lockutils [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.286073] env[61906]: DEBUG oslo_concurrency.lockutils [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.286255] env[61906]: DEBUG oslo_concurrency.lockutils [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.288385] env[61906]: INFO nova.compute.manager [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Terminating instance [ 1114.290244] env[61906]: DEBUG nova.compute.manager [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1114.290484] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1114.291476] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b4f896-f204-46de-bf39-04832650e663 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.300680] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1114.300937] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dcb651cf-c29d-4d8c-949a-7b2d038a8b6d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.307057] env[61906]: DEBUG oslo_vmware.api [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1114.307057] env[61906]: value = "task-1357150" [ 1114.307057] env[61906]: _type = "Task" [ 1114.307057] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.314738] env[61906]: DEBUG oslo_vmware.api [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357150, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.316349] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Acquiring lock "refresh_cache-9f78bfe4-f8f2-44c5-a704-5e3763822211" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.316483] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Acquired lock "refresh_cache-9f78bfe4-f8f2-44c5-a704-5e3763822211" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.316626] env[61906]: DEBUG nova.network.neutron [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1114.423763] env[61906]: DEBUG nova.network.neutron [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Successfully created port: 14e03727-7e3e-434c-8547-4aea5faab8ff {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1114.587640] env[61906]: DEBUG nova.compute.manager [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1114.672221] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 91dc442c-0272-419d-8ef9-06e279a3ea61 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1114.672351] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance d8c1fc04-d5ac-4c78-95e1-e135b22819c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1114.672477] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance fc0e81b8-86b8-4396-afdb-fe9184815fc0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1114.672602] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 9f78bfe4-f8f2-44c5-a704-5e3763822211 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1114.672713] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 43530cb7-293f-4b50-bf9b-e8ff1b047484 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1114.673014] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1114.673076] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1114.739596] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5d4ce1-dd0e-4f68-ae22-7459e762887b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.747058] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8f6f9a-e821-4236-9c63-e2b086828cb1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.777508] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdb07b0-aded-4ca5-a499-12b1883e2279 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.784767] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b2734a-6051-4432-be83-767c1e3a10f7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.797828] env[61906]: DEBUG nova.compute.provider_tree [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.816736] env[61906]: DEBUG oslo_vmware.api [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357150, 'name': PowerOffVM_Task, 'duration_secs': 0.179972} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.817009] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1114.817195] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1114.817444] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8482c213-ba87-46a6-a013-08ae175e9d31 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.848957] env[61906]: DEBUG nova.network.neutron [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1114.982310] env[61906]: DEBUG nova.network.neutron [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Updating instance_info_cache with network_info: [{"id": "1e0e008f-84b1-42c4-a2d8-9663a0bb4b63", "address": "fa:16:3e:fa:ee:e0", "network": {"id": "1017a833-939f-423d-b0dd-bddab58d0dd2", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1532350379-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fef135be23341e9a47f5b22370013a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e0e008f-84", "ovs_interfaceid": "1e0e008f-84b1-42c4-a2d8-9663a0bb4b63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.249940] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1115.250174] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1115.250366] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleting the datastore file [datastore2] fc0e81b8-86b8-4396-afdb-fe9184815fc0 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1115.250671] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f1e2eced-c16d-4f9d-8588-23cfe296b141 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.257513] env[61906]: DEBUG oslo_vmware.api [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1115.257513] env[61906]: value = "task-1357152" [ 1115.257513] env[61906]: _type = "Task" [ 1115.257513] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.265341] env[61906]: DEBUG oslo_vmware.api [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357152, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.300676] env[61906]: DEBUG nova.scheduler.client.report [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1115.486036] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Releasing lock "refresh_cache-9f78bfe4-f8f2-44c5-a704-5e3763822211" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.486036] env[61906]: DEBUG nova.compute.manager [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Instance network_info: |[{"id": "1e0e008f-84b1-42c4-a2d8-9663a0bb4b63", "address": "fa:16:3e:fa:ee:e0", "network": {"id": "1017a833-939f-423d-b0dd-bddab58d0dd2", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1532350379-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fef135be23341e9a47f5b22370013a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e0e008f-84", "ovs_interfaceid": "1e0e008f-84b1-42c4-a2d8-9663a0bb4b63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1115.486036] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:ee:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ffb921-796a-40fe-9662-d3fc01547dcb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1e0e008f-84b1-42c4-a2d8-9663a0bb4b63', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1115.493550] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Creating folder: Project (7fef135be23341e9a47f5b22370013a5). Parent ref: group-v288914. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1115.493831] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aff0cd6a-a4ac-46ed-b774-6e0d227e77ca {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.504673] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Created folder: Project (7fef135be23341e9a47f5b22370013a5) in parent group-v288914. [ 1115.504915] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Creating folder: Instances. Parent ref: group-v289052. {{(pid=61906) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1115.505183] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0e5f1b94-f7c7-49c5-a199-1a8aa43dd065 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.514091] env[61906]: INFO nova.virt.vmwareapi.vm_util [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Created folder: Instances in parent group-v289052. [ 1115.514314] env[61906]: DEBUG oslo.service.loopingcall [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1115.514494] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1115.514680] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f6b084a-7022-490f-8f23-788c97ffc4a8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.532706] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1115.532706] env[61906]: value = "task-1357155" [ 1115.532706] env[61906]: _type = "Task" [ 1115.532706] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.539641] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357155, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.597488] env[61906]: DEBUG nova.compute.manager [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1115.622768] env[61906]: DEBUG nova.virt.hardware [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1115.623029] env[61906]: DEBUG nova.virt.hardware [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1115.623202] env[61906]: DEBUG nova.virt.hardware [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1115.623389] env[61906]: DEBUG nova.virt.hardware [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1115.623542] env[61906]: DEBUG nova.virt.hardware [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1115.623782] env[61906]: DEBUG nova.virt.hardware [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1115.624033] env[61906]: DEBUG nova.virt.hardware [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1115.624210] env[61906]: DEBUG nova.virt.hardware [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1115.624385] env[61906]: DEBUG nova.virt.hardware [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1115.624555] env[61906]: DEBUG nova.virt.hardware [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1115.624735] env[61906]: DEBUG nova.virt.hardware [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1115.625592] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a0b7ee6-c39c-4337-8304-cf5189cb5dfc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.634040] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe69ac1f-0368-4fb8-9cdb-25a30ed0c146 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.768851] env[61906]: DEBUG oslo_vmware.api [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357152, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162521} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.770738] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1115.771101] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1115.771412] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1115.771716] env[61906]: INFO nova.compute.manager [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Took 1.48 seconds to destroy the instance on the hypervisor. [ 1115.772143] env[61906]: DEBUG oslo.service.loopingcall [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1115.773850] env[61906]: DEBUG nova.compute.manager [req-bb45aa1c-720f-4b50-a003-6f8a7e6fe43a req-c9c42efb-bfcb-4a42-96cc-248f0e338dd2 service nova] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Received event network-changed-1e0e008f-84b1-42c4-a2d8-9663a0bb4b63 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1115.774453] env[61906]: DEBUG nova.compute.manager [req-bb45aa1c-720f-4b50-a003-6f8a7e6fe43a req-c9c42efb-bfcb-4a42-96cc-248f0e338dd2 service nova] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Refreshing instance network info cache due to event network-changed-1e0e008f-84b1-42c4-a2d8-9663a0bb4b63. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1115.774453] env[61906]: DEBUG oslo_concurrency.lockutils [req-bb45aa1c-720f-4b50-a003-6f8a7e6fe43a req-c9c42efb-bfcb-4a42-96cc-248f0e338dd2 service nova] Acquiring lock "refresh_cache-9f78bfe4-f8f2-44c5-a704-5e3763822211" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.774647] env[61906]: DEBUG oslo_concurrency.lockutils [req-bb45aa1c-720f-4b50-a003-6f8a7e6fe43a req-c9c42efb-bfcb-4a42-96cc-248f0e338dd2 service nova] Acquired lock "refresh_cache-9f78bfe4-f8f2-44c5-a704-5e3763822211" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.774927] env[61906]: DEBUG nova.network.neutron [req-bb45aa1c-720f-4b50-a003-6f8a7e6fe43a req-c9c42efb-bfcb-4a42-96cc-248f0e338dd2 service nova] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Refreshing network info cache for port 1e0e008f-84b1-42c4-a2d8-9663a0bb4b63 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1115.776758] env[61906]: DEBUG nova.compute.manager [-] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1115.776903] env[61906]: DEBUG nova.network.neutron [-] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1115.804923] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1115.805161] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.165s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.989032] env[61906]: DEBUG nova.network.neutron [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Successfully updated port: 14e03727-7e3e-434c-8547-4aea5faab8ff {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1116.043178] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357155, 'name': CreateVM_Task, 'duration_secs': 0.353521} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.043360] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1116.044016] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.044197] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.044557] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1116.044817] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6252dba-6e02-4274-af7f-3e84410583b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.049149] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Waiting for the task: (returnval){ [ 1116.049149] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c2bdd7-f657-52a0-2261-949434a393c8" [ 1116.049149] env[61906]: _type = "Task" [ 1116.049149] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.056719] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c2bdd7-f657-52a0-2261-949434a393c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.074982] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.075167] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1116.075298] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Rebuilding the list of instances to heal {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1116.289568] env[61906]: DEBUG nova.compute.manager [req-ddd71aa5-b4cf-4782-899f-14716f505b1c req-e1bb7c0b-f4d2-4705-a345-c4118cb291f5 service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Received event network-vif-deleted-da2e07e3-7453-4d91-88f7-26247bd7ea9d {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1116.289568] env[61906]: INFO nova.compute.manager [req-ddd71aa5-b4cf-4782-899f-14716f505b1c req-e1bb7c0b-f4d2-4705-a345-c4118cb291f5 service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Neutron deleted interface da2e07e3-7453-4d91-88f7-26247bd7ea9d; detaching it from the instance and deleting it from the info cache [ 1116.289743] env[61906]: DEBUG nova.network.neutron [req-ddd71aa5-b4cf-4782-899f-14716f505b1c req-e1bb7c0b-f4d2-4705-a345-c4118cb291f5 service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.497585] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "refresh_cache-43530cb7-293f-4b50-bf9b-e8ff1b047484" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.497585] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "refresh_cache-43530cb7-293f-4b50-bf9b-e8ff1b047484" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.497585] env[61906]: DEBUG nova.network.neutron [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1116.513164] env[61906]: DEBUG nova.network.neutron [req-bb45aa1c-720f-4b50-a003-6f8a7e6fe43a req-c9c42efb-bfcb-4a42-96cc-248f0e338dd2 service nova] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Updated VIF entry in instance network info cache for port 1e0e008f-84b1-42c4-a2d8-9663a0bb4b63. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1116.513517] env[61906]: DEBUG nova.network.neutron [req-bb45aa1c-720f-4b50-a003-6f8a7e6fe43a req-c9c42efb-bfcb-4a42-96cc-248f0e338dd2 service nova] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Updating instance_info_cache with network_info: [{"id": "1e0e008f-84b1-42c4-a2d8-9663a0bb4b63", "address": "fa:16:3e:fa:ee:e0", "network": {"id": "1017a833-939f-423d-b0dd-bddab58d0dd2", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1532350379-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fef135be23341e9a47f5b22370013a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e0e008f-84", "ovs_interfaceid": "1e0e008f-84b1-42c4-a2d8-9663a0bb4b63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.560569] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c2bdd7-f657-52a0-2261-949434a393c8, 'name': SearchDatastore_Task, 'duration_secs': 0.008778} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.560917] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.561193] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1116.561436] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.561592] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.561777] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1116.562090] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eecbd558-9b26-46e3-933c-8da8da425a72 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.569642] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1116.569827] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1116.570553] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-664cf713-bd1f-48cb-9af5-01eb48ec5c8f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.575181] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Waiting for the task: (returnval){ [ 1116.575181] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525b7614-adc7-6bdd-3746-695daf821bda" [ 1116.575181] env[61906]: _type = "Task" [ 1116.575181] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.579249] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Skipping network cache update for instance because it is being deleted. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 1116.579430] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1116.579506] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Skipping network cache update for instance because it is Building. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1116.585980] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]525b7614-adc7-6bdd-3746-695daf821bda, 'name': SearchDatastore_Task, 'duration_secs': 0.008111} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.586665] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4fdd95e-16b2-4244-9580-1b9baa72b02f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.591456] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Waiting for the task: (returnval){ [ 1116.591456] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a83f38-d305-150c-d029-e88ae5b5c6e5" [ 1116.591456] env[61906]: _type = "Task" [ 1116.591456] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.599016] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a83f38-d305-150c-d029-e88ae5b5c6e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.605845] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.605983] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquired lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.606151] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Forcefully refreshing network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1116.606307] env[61906]: DEBUG nova.objects.instance [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lazy-loading 'info_cache' on Instance uuid d8c1fc04-d5ac-4c78-95e1-e135b22819c0 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.737507] env[61906]: DEBUG nova.network.neutron [-] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.792497] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4971d48e-5c30-42e1-96af-b9b15a1285a7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.802147] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999b9d83-9cbe-4248-9147-7d2b224588c3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.827819] env[61906]: DEBUG nova.compute.manager [req-ddd71aa5-b4cf-4782-899f-14716f505b1c req-e1bb7c0b-f4d2-4705-a345-c4118cb291f5 service nova] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Detach interface failed, port_id=da2e07e3-7453-4d91-88f7-26247bd7ea9d, reason: Instance fc0e81b8-86b8-4396-afdb-fe9184815fc0 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1117.015698] env[61906]: DEBUG oslo_concurrency.lockutils [req-bb45aa1c-720f-4b50-a003-6f8a7e6fe43a req-c9c42efb-bfcb-4a42-96cc-248f0e338dd2 service nova] Releasing lock "refresh_cache-9f78bfe4-f8f2-44c5-a704-5e3763822211" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.027615] env[61906]: DEBUG nova.network.neutron [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1117.101261] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52a83f38-d305-150c-d029-e88ae5b5c6e5, 'name': SearchDatastore_Task, 'duration_secs': 0.008644} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.101575] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.101832] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 9f78bfe4-f8f2-44c5-a704-5e3763822211/9f78bfe4-f8f2-44c5-a704-5e3763822211.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1117.102101] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8fc89afb-f9f9-46d3-95be-a146db49ef82 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.109906] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Waiting for the task: (returnval){ [ 1117.109906] env[61906]: value = "task-1357156" [ 1117.109906] env[61906]: _type = "Task" [ 1117.109906] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.117214] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357156, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.165294] env[61906]: DEBUG nova.network.neutron [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Updating instance_info_cache with network_info: [{"id": "14e03727-7e3e-434c-8547-4aea5faab8ff", "address": "fa:16:3e:11:00:ca", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14e03727-7e", "ovs_interfaceid": "14e03727-7e3e-434c-8547-4aea5faab8ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.239481] env[61906]: INFO nova.compute.manager [-] [instance: fc0e81b8-86b8-4396-afdb-fe9184815fc0] Took 1.46 seconds to deallocate network for instance. [ 1117.621898] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357156, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477016} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.622260] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 9f78bfe4-f8f2-44c5-a704-5e3763822211/9f78bfe4-f8f2-44c5-a704-5e3763822211.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1117.622381] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1117.622627] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-821529a7-3d66-4155-8d7f-e52629649bba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.628495] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Waiting for the task: (returnval){ [ 1117.628495] env[61906]: value = "task-1357157" [ 1117.628495] env[61906]: _type = "Task" [ 1117.628495] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.635678] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357157, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.668414] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "refresh_cache-43530cb7-293f-4b50-bf9b-e8ff1b047484" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.668685] env[61906]: DEBUG nova.compute.manager [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Instance network_info: |[{"id": "14e03727-7e3e-434c-8547-4aea5faab8ff", "address": "fa:16:3e:11:00:ca", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14e03727-7e", "ovs_interfaceid": "14e03727-7e3e-434c-8547-4aea5faab8ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1117.669112] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:00:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1895250-76cc-41f7-b7f8-2e5679494607', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '14e03727-7e3e-434c-8547-4aea5faab8ff', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1117.676359] env[61906]: DEBUG oslo.service.loopingcall [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1117.676560] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1117.676776] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-29e0e9a6-e40a-4ed9-92f5-ca75c27d13c0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.695057] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1117.695057] env[61906]: value = "task-1357158" [ 1117.695057] env[61906]: _type = "Task" [ 1117.695057] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.702115] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357158, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.747974] env[61906]: DEBUG oslo_concurrency.lockutils [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.748280] env[61906]: DEBUG oslo_concurrency.lockutils [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.748487] env[61906]: DEBUG nova.objects.instance [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lazy-loading 'resources' on Instance uuid fc0e81b8-86b8-4396-afdb-fe9184815fc0 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.798665] env[61906]: DEBUG nova.compute.manager [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Received event network-vif-plugged-14e03727-7e3e-434c-8547-4aea5faab8ff {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1117.798856] env[61906]: DEBUG oslo_concurrency.lockutils [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] Acquiring lock "43530cb7-293f-4b50-bf9b-e8ff1b047484-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.799091] env[61906]: DEBUG oslo_concurrency.lockutils [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] Lock "43530cb7-293f-4b50-bf9b-e8ff1b047484-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.799266] env[61906]: DEBUG oslo_concurrency.lockutils [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] Lock "43530cb7-293f-4b50-bf9b-e8ff1b047484-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.799438] env[61906]: DEBUG nova.compute.manager [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] No waiting events found dispatching network-vif-plugged-14e03727-7e3e-434c-8547-4aea5faab8ff {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1117.799679] env[61906]: WARNING nova.compute.manager [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Received unexpected event network-vif-plugged-14e03727-7e3e-434c-8547-4aea5faab8ff for instance with vm_state building and task_state spawning. [ 1117.799863] env[61906]: DEBUG nova.compute.manager [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Received event network-changed-14e03727-7e3e-434c-8547-4aea5faab8ff {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1117.800032] env[61906]: DEBUG nova.compute.manager [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Refreshing instance network info cache due to event network-changed-14e03727-7e3e-434c-8547-4aea5faab8ff. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1117.800228] env[61906]: DEBUG oslo_concurrency.lockutils [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] Acquiring lock "refresh_cache-43530cb7-293f-4b50-bf9b-e8ff1b047484" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.800370] env[61906]: DEBUG oslo_concurrency.lockutils [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] Acquired lock "refresh_cache-43530cb7-293f-4b50-bf9b-e8ff1b047484" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.800544] env[61906]: DEBUG nova.network.neutron [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Refreshing network info cache for port 14e03727-7e3e-434c-8547-4aea5faab8ff {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1118.140115] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357157, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059997} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.140415] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1118.141223] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-737a4a02-0144-4527-a275-f9f9f20515b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.163452] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 9f78bfe4-f8f2-44c5-a704-5e3763822211/9f78bfe4-f8f2-44c5-a704-5e3763822211.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1118.164644] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2edc962-3960-4ba0-ad78-c7b7cd165e39 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.182317] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Waiting for the task: (returnval){ [ 1118.182317] env[61906]: value = "task-1357159" [ 1118.182317] env[61906]: _type = "Task" [ 1118.182317] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.191189] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357159, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.203100] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357158, 'name': CreateVM_Task, 'duration_secs': 0.272119} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.203274] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1118.203946] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.204156] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.204483] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1118.204741] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c11328bf-178d-4c64-8087-fa03725e3f61 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.208942] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1118.208942] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5288268a-a6a0-d001-0030-abddf9245e93" [ 1118.208942] env[61906]: _type = "Task" [ 1118.208942] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.218631] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5288268a-a6a0-d001-0030-abddf9245e93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.336827] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4b4313-534b-4c6b-8268-edfb2a91cdae {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.346396] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d03a68-6184-4150-a314-fda9e0509407 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.378600] env[61906]: DEBUG nova.network.neutron [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updating instance_info_cache with network_info: [{"id": "3c075a5e-3c44-45b5-bada-620921c0a005", "address": "fa:16:3e:39:60:50", "network": {"id": "f257098c-13ca-45d2-95a2-b74f6a5be9cc", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-455637111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f3bb9ce83b3430bb9202d3c72e77b4c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8edfde4-5a99-4745-956d-04da82ab1b85", "external-id": "nsx-vlan-transportzone-519", "segmentation_id": 519, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c075a5e-3c", "ovs_interfaceid": "3c075a5e-3c44-45b5-bada-620921c0a005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.380302] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545c99b5-583f-41ba-a70f-7066b301838f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.388745] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b499dc16-4201-4eae-af48-983d133929eb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.404645] env[61906]: DEBUG nova.compute.provider_tree [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.515654] env[61906]: DEBUG nova.network.neutron [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Updated VIF entry in instance network info cache for port 14e03727-7e3e-434c-8547-4aea5faab8ff. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1118.516023] env[61906]: DEBUG nova.network.neutron [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Updating instance_info_cache with network_info: [{"id": "14e03727-7e3e-434c-8547-4aea5faab8ff", "address": "fa:16:3e:11:00:ca", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14e03727-7e", "ovs_interfaceid": "14e03727-7e3e-434c-8547-4aea5faab8ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.692104] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357159, 'name': ReconfigVM_Task, 'duration_secs': 0.268302} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.692473] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 9f78bfe4-f8f2-44c5-a704-5e3763822211/9f78bfe4-f8f2-44c5-a704-5e3763822211.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1118.693150] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78a8b3b5-9836-4de2-8942-192a35f55cc2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.699373] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Waiting for the task: (returnval){ [ 1118.699373] env[61906]: value = "task-1357160" [ 1118.699373] env[61906]: _type = "Task" [ 1118.699373] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.706501] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357160, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.716775] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]5288268a-a6a0-d001-0030-abddf9245e93, 'name': SearchDatastore_Task, 'duration_secs': 0.008989} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.717066] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.717305] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1118.717552] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.717698] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.717943] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1118.718268] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19572804-daf7-4cdd-aa4f-793ba764bc8e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.726033] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1118.726301] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1118.727088] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f887ac2-36a7-4df7-a625-eb3e044a56c4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.732226] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1118.732226] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fa530e-e68e-1d37-7d3d-a4630da15e51" [ 1118.732226] env[61906]: _type = "Task" [ 1118.732226] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.738935] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fa530e-e68e-1d37-7d3d-a4630da15e51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.883989] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Releasing lock "refresh_cache-d8c1fc04-d5ac-4c78-95e1-e135b22819c0" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.884264] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updated the network info_cache for instance {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1118.884492] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1118.884660] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1118.884810] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1118.909403] env[61906]: DEBUG nova.scheduler.client.report [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1119.018286] env[61906]: DEBUG oslo_concurrency.lockutils [req-2f54a4ee-09e7-41f2-8401-701250c50894 req-0b7c4a55-d9c8-4df5-8867-a359147c2727 service nova] Releasing lock "refresh_cache-43530cb7-293f-4b50-bf9b-e8ff1b047484" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.209861] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357160, 'name': Rename_Task, 'duration_secs': 0.170691} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.210104] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1119.210360] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8791093-2f65-484c-b257-08ad6b627e13 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.216054] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Waiting for the task: (returnval){ [ 1119.216054] env[61906]: value = "task-1357161" [ 1119.216054] env[61906]: _type = "Task" [ 1119.216054] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.222929] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357161, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.240856] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52fa530e-e68e-1d37-7d3d-a4630da15e51, 'name': SearchDatastore_Task, 'duration_secs': 0.008213} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.241574] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93bb96d5-f4a1-4133-9475-7e4857c72441 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.246172] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1119.246172] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528ba622-6a4c-4cd0-1af2-dcee52486334" [ 1119.246172] env[61906]: _type = "Task" [ 1119.246172] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.253433] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528ba622-6a4c-4cd0-1af2-dcee52486334, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.388561] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Getting list of instances from cluster (obj){ [ 1119.388561] env[61906]: value = "domain-c8" [ 1119.388561] env[61906]: _type = "ClusterComputeResource" [ 1119.388561] env[61906]: } {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1119.389704] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01cccc18-e5d0-44e9-bc05-2044892eadbd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.401922] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Got total of 4 instances {{(pid=61906) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1119.402146] env[61906]: WARNING nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] While synchronizing instance power states, found 5 instances in the database and 4 instances on the hypervisor. [ 1119.402304] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Triggering sync for uuid d8c1fc04-d5ac-4c78-95e1-e135b22819c0 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1119.402509] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Triggering sync for uuid 91dc442c-0272-419d-8ef9-06e279a3ea61 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1119.402662] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Triggering sync for uuid fc0e81b8-86b8-4396-afdb-fe9184815fc0 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1119.402819] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Triggering sync for uuid 9f78bfe4-f8f2-44c5-a704-5e3763822211 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1119.402967] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Triggering sync for uuid 43530cb7-293f-4b50-bf9b-e8ff1b047484 {{(pid=61906) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1119.403318] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.403537] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.403806] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "91dc442c-0272-419d-8ef9-06e279a3ea61" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.403998] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "91dc442c-0272-419d-8ef9-06e279a3ea61" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.404253] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.404467] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "9f78bfe4-f8f2-44c5-a704-5e3763822211" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.404678] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "43530cb7-293f-4b50-bf9b-e8ff1b047484" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.405498] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34feb9ed-3728-43b1-8d0c-aaf083f50d67 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.408415] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fcbdb76-9e12-463a-b88a-1ce1e3cff600 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.414705] env[61906]: DEBUG oslo_concurrency.lockutils [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.666s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.438487] env[61906]: INFO nova.scheduler.client.report [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleted allocations for instance fc0e81b8-86b8-4396-afdb-fe9184815fc0 [ 1119.726427] env[61906]: DEBUG oslo_vmware.api [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357161, 'name': PowerOnVM_Task, 'duration_secs': 0.44642} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.726728] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1119.726803] env[61906]: INFO nova.compute.manager [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Took 6.29 seconds to spawn the instance on the hypervisor. [ 1119.726989] env[61906]: DEBUG nova.compute.manager [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1119.727776] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29f30bb-db04-4fc0-a1b3-d5313d1fe9a2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.754621] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]528ba622-6a4c-4cd0-1af2-dcee52486334, 'name': SearchDatastore_Task, 'duration_secs': 0.009411} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.754872] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.755152] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 43530cb7-293f-4b50-bf9b-e8ff1b047484/43530cb7-293f-4b50-bf9b-e8ff1b047484.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1119.755400] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b8cfe02e-46b1-49ba-9fff-008db138acd9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.760856] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1119.760856] env[61906]: value = "task-1357162" [ 1119.760856] env[61906]: _type = "Task" [ 1119.760856] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.768282] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357162, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.922153] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.518s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.922610] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "91dc442c-0272-419d-8ef9-06e279a3ea61" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.518s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.946650] env[61906]: DEBUG oslo_concurrency.lockutils [None req-adea5a72-0561-46f9-8fc8-c21d5d612f37 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.661s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.948066] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.543s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.948307] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-30a1eafa-8d3c-4540-b9e2-e196292968ba {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.958199] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99003e3-a08f-44f9-a9da-c123ec8992fd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.242953] env[61906]: INFO nova.compute.manager [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Took 11.00 seconds to build instance. [ 1120.271656] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357162, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453211} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.271998] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore2] 43530cb7-293f-4b50-bf9b-e8ff1b047484/43530cb7-293f-4b50-bf9b-e8ff1b047484.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1120.272305] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1120.272620] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a8b9c879-460e-4aae-b64a-dc719bfa2f7e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.279222] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1120.279222] env[61906]: value = "task-1357163" [ 1120.279222] env[61906]: _type = "Task" [ 1120.279222] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.289246] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357163, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.485749] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "fc0e81b8-86b8-4396-afdb-fe9184815fc0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.538s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.746388] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2ed2bce7-b76f-4a5a-ad20-fb2e2a58556c tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Lock "9f78bfe4-f8f2-44c5-a704-5e3763822211" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.511s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.746699] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "9f78bfe4-f8f2-44c5-a704-5e3763822211" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.342s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.746871] env[61906]: INFO nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] During sync_power_state the instance has a pending task (spawning). Skip. [ 1120.747043] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "9f78bfe4-f8f2-44c5-a704-5e3763822211" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.788787] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357163, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103345} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.789091] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1120.789971] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffbbeec-022f-44fc-a916-4b81bf7343bd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.813561] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 43530cb7-293f-4b50-bf9b-e8ff1b047484/43530cb7-293f-4b50-bf9b-e8ff1b047484.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1120.813877] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc235ea3-b5fb-4263-a0fb-36ed79ba9062 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.833520] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1120.833520] env[61906]: value = "task-1357164" [ 1120.833520] env[61906]: _type = "Task" [ 1120.833520] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.844041] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357164, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.331042] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.331292] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.343348] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357164, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.432073] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Acquiring lock "9f78bfe4-f8f2-44c5-a704-5e3763822211" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.432387] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Lock "9f78bfe4-f8f2-44c5-a704-5e3763822211" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.432681] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Acquiring lock "9f78bfe4-f8f2-44c5-a704-5e3763822211-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.432876] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Lock "9f78bfe4-f8f2-44c5-a704-5e3763822211-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.433108] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Lock "9f78bfe4-f8f2-44c5-a704-5e3763822211-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.435267] env[61906]: INFO nova.compute.manager [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Terminating instance [ 1121.437046] env[61906]: DEBUG nova.compute.manager [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1121.437249] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1121.438090] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e05f50-dade-4130-8055-2b1c840be9cd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.445532] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1121.445744] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5090ea80-43a8-4ca0-ba93-ae4225f3f6d2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.451686] env[61906]: DEBUG oslo_vmware.api [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Waiting for the task: (returnval){ [ 1121.451686] env[61906]: value = "task-1357165" [ 1121.451686] env[61906]: _type = "Task" [ 1121.451686] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.458664] env[61906]: DEBUG oslo_vmware.api [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357165, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.833856] env[61906]: DEBUG nova.compute.manager [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1121.845825] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357164, 'name': ReconfigVM_Task, 'duration_secs': 0.832196} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.846145] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 43530cb7-293f-4b50-bf9b-e8ff1b047484/43530cb7-293f-4b50-bf9b-e8ff1b047484.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1121.846845] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-746607c3-f455-450f-9d8e-9e22c5c9c285 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.854037] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1121.854037] env[61906]: value = "task-1357166" [ 1121.854037] env[61906]: _type = "Task" [ 1121.854037] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.861144] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357166, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.961747] env[61906]: DEBUG oslo_vmware.api [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357165, 'name': PowerOffVM_Task, 'duration_secs': 0.19201} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.962155] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1121.962355] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1121.962643] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4a5d442f-2519-475f-9b60-8f0ed28b81b8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.046152] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1122.046391] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1122.046581] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Deleting the datastore file [datastore2] 9f78bfe4-f8f2-44c5-a704-5e3763822211 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1122.046866] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1001232-fd75-48cc-b8c5-d4e2c25b6a1c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.053216] env[61906]: DEBUG oslo_vmware.api [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Waiting for the task: (returnval){ [ 1122.053216] env[61906]: value = "task-1357168" [ 1122.053216] env[61906]: _type = "Task" [ 1122.053216] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.060719] env[61906]: DEBUG oslo_vmware.api [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357168, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.360060] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.360060] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.361553] env[61906]: INFO nova.compute.claims [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1122.367116] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357166, 'name': Rename_Task, 'duration_secs': 0.138449} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.367581] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1122.367829] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9fa02d50-3207-4033-bc38-a39477f19caa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.374032] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1122.374032] env[61906]: value = "task-1357169" [ 1122.374032] env[61906]: _type = "Task" [ 1122.374032] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.381347] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357169, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.562803] env[61906]: DEBUG oslo_vmware.api [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Task: {'id': task-1357168, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154688} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.563107] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1122.563309] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1122.563490] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1122.563666] env[61906]: INFO nova.compute.manager [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1122.563912] env[61906]: DEBUG oslo.service.loopingcall [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1122.564131] env[61906]: DEBUG nova.compute.manager [-] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1122.564227] env[61906]: DEBUG nova.network.neutron [-] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1122.795680] env[61906]: DEBUG nova.compute.manager [req-36c96220-8764-43c8-a018-7ffc95eff6b3 req-0f3e07b2-edcd-4475-9038-8565de17eb98 service nova] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Received event network-vif-deleted-1e0e008f-84b1-42c4-a2d8-9663a0bb4b63 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1122.796014] env[61906]: INFO nova.compute.manager [req-36c96220-8764-43c8-a018-7ffc95eff6b3 req-0f3e07b2-edcd-4475-9038-8565de17eb98 service nova] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Neutron deleted interface 1e0e008f-84b1-42c4-a2d8-9663a0bb4b63; detaching it from the instance and deleting it from the info cache [ 1122.796222] env[61906]: DEBUG nova.network.neutron [req-36c96220-8764-43c8-a018-7ffc95eff6b3 req-0f3e07b2-edcd-4475-9038-8565de17eb98 service nova] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.884156] env[61906]: DEBUG oslo_vmware.api [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357169, 'name': PowerOnVM_Task, 'duration_secs': 0.482907} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.884464] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1122.884636] env[61906]: INFO nova.compute.manager [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Took 7.29 seconds to spawn the instance on the hypervisor. [ 1122.884817] env[61906]: DEBUG nova.compute.manager [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1122.885668] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d546ecab-beb1-43a7-a3cc-51c45525e3c8 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.274464] env[61906]: DEBUG nova.network.neutron [-] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.298506] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ede235b5-52d4-4748-9f80-00f040bef4d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.308274] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4547cf0f-b23f-4b9d-b272-77da69617e8e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.334470] env[61906]: DEBUG nova.compute.manager [req-36c96220-8764-43c8-a018-7ffc95eff6b3 req-0f3e07b2-edcd-4475-9038-8565de17eb98 service nova] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Detach interface failed, port_id=1e0e008f-84b1-42c4-a2d8-9663a0bb4b63, reason: Instance 9f78bfe4-f8f2-44c5-a704-5e3763822211 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1123.406087] env[61906]: INFO nova.compute.manager [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Took 13.59 seconds to build instance. [ 1123.443743] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89987bc6-bf08-480d-9c4f-e5ab28fd06d1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.451276] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da034f60-f912-4c0a-9987-9b4362a1eb5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.480334] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44458b00-a5d2-400c-a3fb-bf475f05128f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.488334] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1de73e2-79c6-4a3c-83a8-40ffa2a3c623 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.501950] env[61906]: DEBUG nova.compute.provider_tree [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1123.659875] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.660183] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.660734] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.660734] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.660880] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.663047] env[61906]: INFO nova.compute.manager [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Terminating instance [ 1123.664961] env[61906]: DEBUG nova.compute.manager [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1123.665171] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1123.666045] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e81238-0235-4fa7-b9c8-6172ab410d47 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.674137] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1123.674375] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6312431-1d22-4b84-9b0b-2378f661eb17 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.680142] env[61906]: DEBUG oslo_vmware.api [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1123.680142] env[61906]: value = "task-1357170" [ 1123.680142] env[61906]: _type = "Task" [ 1123.680142] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.687796] env[61906]: DEBUG oslo_vmware.api [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357170, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.778892] env[61906]: INFO nova.compute.manager [-] [instance: 9f78bfe4-f8f2-44c5-a704-5e3763822211] Took 1.21 seconds to deallocate network for instance. [ 1123.908058] env[61906]: DEBUG oslo_concurrency.lockutils [None req-19b93b6f-cbb8-4106-a680-2ee3dfd8e3b2 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "43530cb7-293f-4b50-bf9b-e8ff1b047484" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.100s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.908351] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "43530cb7-293f-4b50-bf9b-e8ff1b047484" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.504s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.908519] env[61906]: INFO nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] During sync_power_state the instance has a pending task (spawning). Skip. [ 1123.908707] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "43530cb7-293f-4b50-bf9b-e8ff1b047484" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.005436] env[61906]: DEBUG nova.scheduler.client.report [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1124.189716] env[61906]: DEBUG oslo_vmware.api [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357170, 'name': PowerOffVM_Task, 'duration_secs': 0.180125} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.190009] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1124.190203] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1124.190454] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a324bc36-322f-4470-b13a-365b0e1f8f74 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.268603] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1124.268825] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1124.269019] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleting the datastore file [datastore1] d8c1fc04-d5ac-4c78-95e1-e135b22819c0 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1124.269294] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c1d1c8ec-70c3-4103-82a5-d4c2ae214bec {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.275324] env[61906]: DEBUG oslo_vmware.api [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for the task: (returnval){ [ 1124.275324] env[61906]: value = "task-1357172" [ 1124.275324] env[61906]: _type = "Task" [ 1124.275324] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.283751] env[61906]: DEBUG oslo_vmware.api [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357172, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.284646] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.398545] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "43530cb7-293f-4b50-bf9b-e8ff1b047484" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.398906] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "43530cb7-293f-4b50-bf9b-e8ff1b047484" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.399154] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "43530cb7-293f-4b50-bf9b-e8ff1b047484-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.399386] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "43530cb7-293f-4b50-bf9b-e8ff1b047484-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.399645] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "43530cb7-293f-4b50-bf9b-e8ff1b047484-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.401925] env[61906]: INFO nova.compute.manager [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Terminating instance [ 1124.403755] env[61906]: DEBUG nova.compute.manager [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1124.403966] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1124.404825] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46a60a0-984e-4fef-85b4-6419bf4c545e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.413164] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1124.413405] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a825817-931c-4746-92d7-0f06a834d1b0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.418717] env[61906]: DEBUG oslo_vmware.api [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1124.418717] env[61906]: value = "task-1357173" [ 1124.418717] env[61906]: _type = "Task" [ 1124.418717] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.425894] env[61906]: DEBUG oslo_vmware.api [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357173, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.511033] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.151s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.511575] env[61906]: DEBUG nova.compute.manager [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1124.514406] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.230s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.514648] env[61906]: DEBUG nova.objects.instance [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Lazy-loading 'resources' on Instance uuid 9f78bfe4-f8f2-44c5-a704-5e3763822211 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1124.784987] env[61906]: DEBUG oslo_vmware.api [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Task: {'id': task-1357172, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138322} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.785223] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1124.785413] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1124.785597] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1124.785776] env[61906]: INFO nova.compute.manager [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1124.786045] env[61906]: DEBUG oslo.service.loopingcall [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1124.786257] env[61906]: DEBUG nova.compute.manager [-] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1124.786349] env[61906]: DEBUG nova.network.neutron [-] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1124.928844] env[61906]: DEBUG oslo_vmware.api [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357173, 'name': PowerOffVM_Task, 'duration_secs': 0.161683} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.929193] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1124.929305] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1124.929574] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97b1e3a1-827e-4154-860d-8438b2c62d7e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.989488] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1124.989765] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1124.989962] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleting the datastore file [datastore2] 43530cb7-293f-4b50-bf9b-e8ff1b047484 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1124.990254] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c5314dd7-c27e-4d47-ad4f-ccf2b66c8322 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.996174] env[61906]: DEBUG oslo_vmware.api [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1124.996174] env[61906]: value = "task-1357175" [ 1124.996174] env[61906]: _type = "Task" [ 1124.996174] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.004491] env[61906]: DEBUG oslo_vmware.api [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357175, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.017976] env[61906]: DEBUG nova.compute.utils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1125.019455] env[61906]: DEBUG nova.compute.manager [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1125.019563] env[61906]: DEBUG nova.network.neutron [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1125.105353] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467014c5-daf8-4443-98f5-074a762ed147 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.112745] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31804c2-db99-4a21-ba8b-49426152dffd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.119459] env[61906]: DEBUG nova.policy [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ba1658b4ade413b98834102c39a255c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c6414fa082f49c2a568a772a256a8e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1125.156092] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e069d9-59a2-4cc9-9ad7-da0e41d280a9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.164071] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ac7082-f576-4bf6-95c3-c4957f3b9e32 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.177117] env[61906]: DEBUG nova.compute.provider_tree [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.444074] env[61906]: DEBUG nova.network.neutron [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Successfully created port: c063b2a3-1cf9-458a-a9ba-d3600501cd3e {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1125.451021] env[61906]: DEBUG nova.compute.manager [req-cd46488f-f688-41e7-ae14-76150a5d8e0b req-373c5e96-5eff-4ff9-ab07-5b4efc8c7695 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Received event network-vif-deleted-3c075a5e-3c44-45b5-bada-620921c0a005 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1125.451386] env[61906]: INFO nova.compute.manager [req-cd46488f-f688-41e7-ae14-76150a5d8e0b req-373c5e96-5eff-4ff9-ab07-5b4efc8c7695 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Neutron deleted interface 3c075a5e-3c44-45b5-bada-620921c0a005; detaching it from the instance and deleting it from the info cache [ 1125.451674] env[61906]: DEBUG nova.network.neutron [req-cd46488f-f688-41e7-ae14-76150a5d8e0b req-373c5e96-5eff-4ff9-ab07-5b4efc8c7695 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.506225] env[61906]: DEBUG oslo_vmware.api [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357175, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16011} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.506225] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1125.506225] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1125.506440] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1125.506499] env[61906]: INFO nova.compute.manager [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1125.506735] env[61906]: DEBUG oslo.service.loopingcall [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1125.506923] env[61906]: DEBUG nova.compute.manager [-] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1125.507021] env[61906]: DEBUG nova.network.neutron [-] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1125.525690] env[61906]: DEBUG nova.compute.manager [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1125.681034] env[61906]: DEBUG nova.scheduler.client.report [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1125.930194] env[61906]: DEBUG nova.network.neutron [-] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.955110] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ece42999-a863-420e-b55e-5f833b514e65 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.966097] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5359fda-4c2d-4e98-8a5d-248e26d99823 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.990858] env[61906]: DEBUG nova.compute.manager [req-cd46488f-f688-41e7-ae14-76150a5d8e0b req-373c5e96-5eff-4ff9-ab07-5b4efc8c7695 service nova] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Detach interface failed, port_id=3c075a5e-3c44-45b5-bada-620921c0a005, reason: Instance d8c1fc04-d5ac-4c78-95e1-e135b22819c0 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1126.186011] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.671s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.212485] env[61906]: INFO nova.scheduler.client.report [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Deleted allocations for instance 9f78bfe4-f8f2-44c5-a704-5e3763822211 [ 1126.218367] env[61906]: DEBUG nova.network.neutron [-] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.432944] env[61906]: INFO nova.compute.manager [-] [instance: d8c1fc04-d5ac-4c78-95e1-e135b22819c0] Took 1.65 seconds to deallocate network for instance. [ 1126.536361] env[61906]: DEBUG nova.compute.manager [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1126.561589] env[61906]: DEBUG nova.virt.hardware [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1126.561841] env[61906]: DEBUG nova.virt.hardware [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1126.562009] env[61906]: DEBUG nova.virt.hardware [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1126.562276] env[61906]: DEBUG nova.virt.hardware [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1126.562340] env[61906]: DEBUG nova.virt.hardware [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1126.562494] env[61906]: DEBUG nova.virt.hardware [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1126.562700] env[61906]: DEBUG nova.virt.hardware [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1126.562862] env[61906]: DEBUG nova.virt.hardware [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1126.563141] env[61906]: DEBUG nova.virt.hardware [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1126.563210] env[61906]: DEBUG nova.virt.hardware [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1126.563369] env[61906]: DEBUG nova.virt.hardware [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1126.564235] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b64e9015-98ac-4a43-a2c8-5133d7fe1131 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.572137] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9362d82e-6780-4135-9d0c-1d4cd0461383 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.721142] env[61906]: INFO nova.compute.manager [-] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Took 1.21 seconds to deallocate network for instance. [ 1126.721672] env[61906]: DEBUG oslo_concurrency.lockutils [None req-97085ec7-d3d6-40e4-b35b-217cf93070a5 tempest-ServerAddressesNegativeTestJSON-497675604 tempest-ServerAddressesNegativeTestJSON-497675604-project-member] Lock "9f78bfe4-f8f2-44c5-a704-5e3763822211" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.289s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.939199] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.939466] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.939713] env[61906]: DEBUG nova.objects.instance [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lazy-loading 'resources' on Instance uuid d8c1fc04-d5ac-4c78-95e1-e135b22819c0 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1126.948471] env[61906]: DEBUG nova.compute.manager [req-0e24ce67-f752-4a77-a8f2-fcdf2cbad5ef req-0ea3fc78-6481-4ebe-a8a5-252b706f65a4 service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Received event network-vif-plugged-c063b2a3-1cf9-458a-a9ba-d3600501cd3e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1126.948471] env[61906]: DEBUG oslo_concurrency.lockutils [req-0e24ce67-f752-4a77-a8f2-fcdf2cbad5ef req-0ea3fc78-6481-4ebe-a8a5-252b706f65a4 service nova] Acquiring lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.948471] env[61906]: DEBUG oslo_concurrency.lockutils [req-0e24ce67-f752-4a77-a8f2-fcdf2cbad5ef req-0ea3fc78-6481-4ebe-a8a5-252b706f65a4 service nova] Lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.948471] env[61906]: DEBUG oslo_concurrency.lockutils [req-0e24ce67-f752-4a77-a8f2-fcdf2cbad5ef req-0ea3fc78-6481-4ebe-a8a5-252b706f65a4 service nova] Lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.948471] env[61906]: DEBUG nova.compute.manager [req-0e24ce67-f752-4a77-a8f2-fcdf2cbad5ef req-0ea3fc78-6481-4ebe-a8a5-252b706f65a4 service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] No waiting events found dispatching network-vif-plugged-c063b2a3-1cf9-458a-a9ba-d3600501cd3e {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1126.948471] env[61906]: WARNING nova.compute.manager [req-0e24ce67-f752-4a77-a8f2-fcdf2cbad5ef req-0ea3fc78-6481-4ebe-a8a5-252b706f65a4 service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Received unexpected event network-vif-plugged-c063b2a3-1cf9-458a-a9ba-d3600501cd3e for instance with vm_state building and task_state spawning. [ 1127.117486] env[61906]: DEBUG nova.network.neutron [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Successfully updated port: c063b2a3-1cf9-458a-a9ba-d3600501cd3e {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1127.231492] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.492878] env[61906]: DEBUG nova.compute.manager [req-ffaa4fba-58dd-49cd-992f-1b20d496161e req-bc192b08-469a-4b4c-851c-09072bbd0573 service nova] [instance: 43530cb7-293f-4b50-bf9b-e8ff1b047484] Received event network-vif-deleted-14e03727-7e3e-434c-8547-4aea5faab8ff {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1127.521011] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3dd6ca5-8bc5-4586-ac21-3ed13f5981bc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.529154] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcca7142-216e-4c6d-ae55-1af98584c9b4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.560102] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d980cc9a-0460-444e-8b30-2627172c8c96 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.569020] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0943c714-0df9-4360-9ba2-3053bcf03904 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.581018] env[61906]: DEBUG nova.compute.provider_tree [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1127.620784] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.620935] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.621104] env[61906]: DEBUG nova.network.neutron [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1128.085215] env[61906]: DEBUG nova.scheduler.client.report [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1128.169216] env[61906]: DEBUG nova.network.neutron [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1128.309500] env[61906]: DEBUG nova.network.neutron [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Updating instance_info_cache with network_info: [{"id": "c063b2a3-1cf9-458a-a9ba-d3600501cd3e", "address": "fa:16:3e:75:c9:6b", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc063b2a3-1c", "ovs_interfaceid": "c063b2a3-1cf9-458a-a9ba-d3600501cd3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.590399] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.651s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.592781] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.361s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.593072] env[61906]: DEBUG nova.objects.instance [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lazy-loading 'resources' on Instance uuid 43530cb7-293f-4b50-bf9b-e8ff1b047484 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1128.615160] env[61906]: INFO nova.scheduler.client.report [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Deleted allocations for instance d8c1fc04-d5ac-4c78-95e1-e135b22819c0 [ 1128.812369] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.812704] env[61906]: DEBUG nova.compute.manager [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Instance network_info: |[{"id": "c063b2a3-1cf9-458a-a9ba-d3600501cd3e", "address": "fa:16:3e:75:c9:6b", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc063b2a3-1c", "ovs_interfaceid": "c063b2a3-1cf9-458a-a9ba-d3600501cd3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1128.813148] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:c9:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8233b50c-be21-441a-a357-a29138a95b8b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c063b2a3-1cf9-458a-a9ba-d3600501cd3e', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1128.820639] env[61906]: DEBUG oslo.service.loopingcall [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1128.820874] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1128.821117] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8dbeb302-4071-431a-8015-6fc50d2e7062 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.840253] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1128.840253] env[61906]: value = "task-1357176" [ 1128.840253] env[61906]: _type = "Task" [ 1128.840253] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.848767] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357176, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.973401] env[61906]: DEBUG nova.compute.manager [req-63365fe3-706e-40af-a3a1-404553dadadf req-6e952202-4f52-4bf0-bb01-97f526741232 service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Received event network-changed-c063b2a3-1cf9-458a-a9ba-d3600501cd3e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1128.973638] env[61906]: DEBUG nova.compute.manager [req-63365fe3-706e-40af-a3a1-404553dadadf req-6e952202-4f52-4bf0-bb01-97f526741232 service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Refreshing instance network info cache due to event network-changed-c063b2a3-1cf9-458a-a9ba-d3600501cd3e. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1128.973860] env[61906]: DEBUG oslo_concurrency.lockutils [req-63365fe3-706e-40af-a3a1-404553dadadf req-6e952202-4f52-4bf0-bb01-97f526741232 service nova] Acquiring lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.974014] env[61906]: DEBUG oslo_concurrency.lockutils [req-63365fe3-706e-40af-a3a1-404553dadadf req-6e952202-4f52-4bf0-bb01-97f526741232 service nova] Acquired lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.974186] env[61906]: DEBUG nova.network.neutron [req-63365fe3-706e-40af-a3a1-404553dadadf req-6e952202-4f52-4bf0-bb01-97f526741232 service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Refreshing network info cache for port c063b2a3-1cf9-458a-a9ba-d3600501cd3e {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1129.122828] env[61906]: DEBUG oslo_concurrency.lockutils [None req-1a4dc077-80e9-42e2-9c15-1b2ecb1fb3a0 tempest-AttachVolumeShelveTestJSON-680462611 tempest-AttachVolumeShelveTestJSON-680462611-project-member] Lock "d8c1fc04-d5ac-4c78-95e1-e135b22819c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.462s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.149385] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b1643b-d49d-4335-ae3c-078e84e1ab93 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.156939] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248c3a76-9fb0-428d-9e68-b74366a345a4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.186049] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b25efc93-8280-4afb-9439-eb8373c7c794 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.193536] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569837a1-cb2d-4416-b149-aefeb9b896ed {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.206554] env[61906]: DEBUG nova.compute.provider_tree [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.350151] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357176, 'name': CreateVM_Task, 'duration_secs': 0.288601} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.350320] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1129.351075] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.351279] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.352242] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1129.352242] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae83e2ea-fa66-43ce-b762-19d52105c094 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.356212] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1129.356212] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f855a5-02ff-0c08-9048-6c5691d61109" [ 1129.356212] env[61906]: _type = "Task" [ 1129.356212] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.365143] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f855a5-02ff-0c08-9048-6c5691d61109, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.664936] env[61906]: DEBUG nova.network.neutron [req-63365fe3-706e-40af-a3a1-404553dadadf req-6e952202-4f52-4bf0-bb01-97f526741232 service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Updated VIF entry in instance network info cache for port c063b2a3-1cf9-458a-a9ba-d3600501cd3e. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1129.665329] env[61906]: DEBUG nova.network.neutron [req-63365fe3-706e-40af-a3a1-404553dadadf req-6e952202-4f52-4bf0-bb01-97f526741232 service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Updating instance_info_cache with network_info: [{"id": "c063b2a3-1cf9-458a-a9ba-d3600501cd3e", "address": "fa:16:3e:75:c9:6b", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc063b2a3-1c", "ovs_interfaceid": "c063b2a3-1cf9-458a-a9ba-d3600501cd3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.709630] env[61906]: DEBUG nova.scheduler.client.report [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1129.867906] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52f855a5-02ff-0c08-9048-6c5691d61109, 'name': SearchDatastore_Task, 'duration_secs': 0.009666} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.868240] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.868481] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1129.868719] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.868872] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.869063] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1129.869332] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b4fc3d3-73d1-48bc-8d01-ff49dce94451 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.877016] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1129.877266] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1129.877902] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-879ebf8f-f520-4d0e-8710-0ffa8dbefaeb {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.884091] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1129.884091] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52019836-7346-0132-5bcc-dd1168f3a1f2" [ 1129.884091] env[61906]: _type = "Task" [ 1129.884091] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.889869] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52019836-7346-0132-5bcc-dd1168f3a1f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.168673] env[61906]: DEBUG oslo_concurrency.lockutils [req-63365fe3-706e-40af-a3a1-404553dadadf req-6e952202-4f52-4bf0-bb01-97f526741232 service nova] Releasing lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.214854] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.622s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.239045] env[61906]: INFO nova.scheduler.client.report [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted allocations for instance 43530cb7-293f-4b50-bf9b-e8ff1b047484 [ 1130.393802] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52019836-7346-0132-5bcc-dd1168f3a1f2, 'name': SearchDatastore_Task, 'duration_secs': 0.008131} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.394574] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db3549da-234c-4642-ba28-2f1d853cbda3 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.400034] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1130.400034] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52833512-36f9-c37f-eed6-eab67e2acd31" [ 1130.400034] env[61906]: _type = "Task" [ 1130.400034] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.407854] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52833512-36f9-c37f-eed6-eab67e2acd31, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.747658] env[61906]: DEBUG oslo_concurrency.lockutils [None req-2dd2f6e8-647d-4ee6-8f7b-9f5aeeaaeb49 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "43530cb7-293f-4b50-bf9b-e8ff1b047484" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.349s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.909866] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52833512-36f9-c37f-eed6-eab67e2acd31, 'name': SearchDatastore_Task, 'duration_secs': 0.010959} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.910171] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.910434] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5/36f6adad-6df4-4e05-ba96-9fbfaaf38ad5.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1130.910718] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ccfdccd8-760f-4fee-9c09-77331815d148 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.916841] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1130.916841] env[61906]: value = "task-1357177" [ 1130.916841] env[61906]: _type = "Task" [ 1130.916841] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.924419] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357177, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.428743] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "660cbb4d-d924-4679-a017-a34c30b88c85" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.429049] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "660cbb4d-d924-4679-a017-a34c30b88c85" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.430535] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357177, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452626} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.431109] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5/36f6adad-6df4-4e05-ba96-9fbfaaf38ad5.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1131.431330] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1131.431567] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f40d956-d2c2-4b13-b80d-7586c49fd178 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.437753] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1131.437753] env[61906]: value = "task-1357179" [ 1131.437753] env[61906]: _type = "Task" [ 1131.437753] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.446389] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357179, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.931032] env[61906]: DEBUG nova.compute.manager [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Starting instance... {{(pid=61906) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1131.948692] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357179, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055103} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.948953] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1131.949752] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf120b5-cef4-4a5b-a137-a6749351649f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.972565] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5/36f6adad-6df4-4e05-ba96-9fbfaaf38ad5.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1131.973478] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29469e79-53de-4de1-bc09-8cc20fd45ef4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.992035] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1131.992035] env[61906]: value = "task-1357180" [ 1131.992035] env[61906]: _type = "Task" [ 1131.992035] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.999666] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357180, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.452842] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.453141] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.454570] env[61906]: INFO nova.compute.claims [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1132.501776] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357180, 'name': ReconfigVM_Task, 'duration_secs': 0.26709} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.502114] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5/36f6adad-6df4-4e05-ba96-9fbfaaf38ad5.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1132.502727] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f41b5d73-daf7-4639-ac15-9808beb501af {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.508722] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1132.508722] env[61906]: value = "task-1357181" [ 1132.508722] env[61906]: _type = "Task" [ 1132.508722] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.516897] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357181, 'name': Rename_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.017762] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357181, 'name': Rename_Task, 'duration_secs': 0.167335} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.018015] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1133.018250] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8ecb5f74-e829-4b40-84d9-1458cca70ba2 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.024105] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1133.024105] env[61906]: value = "task-1357182" [ 1133.024105] env[61906]: _type = "Task" [ 1133.024105] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.030916] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357182, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.515462] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30955bc-3fa6-450d-aac6-385edf87abc4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.523016] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fcbed7e-3ffa-4597-84ec-8e32f3f63432 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.533918] env[61906]: DEBUG oslo_vmware.api [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357182, 'name': PowerOnVM_Task, 'duration_secs': 0.407689} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.557312] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1133.557538] env[61906]: INFO nova.compute.manager [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Took 7.02 seconds to spawn the instance on the hypervisor. [ 1133.557723] env[61906]: DEBUG nova.compute.manager [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1133.558605] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36fc2d89-c249-4c4d-8056-06efa957f923 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.561318] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44adc15-d1bb-49da-beb4-60d1c6a85ef9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.569281] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc17555a-f240-44ee-91be-a469926b9715 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.584705] env[61906]: DEBUG nova.compute.provider_tree [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1134.085783] env[61906]: INFO nova.compute.manager [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Took 11.74 seconds to build instance. [ 1134.087335] env[61906]: DEBUG nova.scheduler.client.report [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1134.593654] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f49a7fa7-c502-4a8d-88f5-c763e90e474b tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.260s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.593654] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.139s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.593654] env[61906]: DEBUG nova.compute.manager [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Start building networks asynchronously for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1134.889970] env[61906]: DEBUG nova.compute.manager [req-62fd6042-4780-437e-a061-75e716522e77 req-7f2474da-f43d-4679-9e45-bfcd5d10434a service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Received event network-changed-c063b2a3-1cf9-458a-a9ba-d3600501cd3e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1134.890189] env[61906]: DEBUG nova.compute.manager [req-62fd6042-4780-437e-a061-75e716522e77 req-7f2474da-f43d-4679-9e45-bfcd5d10434a service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Refreshing instance network info cache due to event network-changed-c063b2a3-1cf9-458a-a9ba-d3600501cd3e. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1134.890414] env[61906]: DEBUG oslo_concurrency.lockutils [req-62fd6042-4780-437e-a061-75e716522e77 req-7f2474da-f43d-4679-9e45-bfcd5d10434a service nova] Acquiring lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.890561] env[61906]: DEBUG oslo_concurrency.lockutils [req-62fd6042-4780-437e-a061-75e716522e77 req-7f2474da-f43d-4679-9e45-bfcd5d10434a service nova] Acquired lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.891142] env[61906]: DEBUG nova.network.neutron [req-62fd6042-4780-437e-a061-75e716522e77 req-7f2474da-f43d-4679-9e45-bfcd5d10434a service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Refreshing network info cache for port c063b2a3-1cf9-458a-a9ba-d3600501cd3e {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1135.097548] env[61906]: DEBUG nova.compute.utils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Using /dev/sd instead of None {{(pid=61906) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1135.098916] env[61906]: DEBUG nova.compute.manager [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Allocating IP information in the background. {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1135.099104] env[61906]: DEBUG nova.network.neutron [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] allocate_for_instance() {{(pid=61906) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1135.155689] env[61906]: DEBUG nova.policy [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e28284ad5ce450597ef12c65c902e1c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ac4f2c9065c46068c7837eb85cd7e20', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61906) authorize /opt/stack/nova/nova/policy.py:201}} [ 1135.602885] env[61906]: DEBUG nova.compute.manager [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Start building block device mappings for instance. {{(pid=61906) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1135.622628] env[61906]: DEBUG nova.network.neutron [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Successfully created port: 6787aaee-7482-4d39-bed2-f88a99b55419 {{(pid=61906) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1135.662045] env[61906]: DEBUG nova.network.neutron [req-62fd6042-4780-437e-a061-75e716522e77 req-7f2474da-f43d-4679-9e45-bfcd5d10434a service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Updated VIF entry in instance network info cache for port c063b2a3-1cf9-458a-a9ba-d3600501cd3e. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1135.662461] env[61906]: DEBUG nova.network.neutron [req-62fd6042-4780-437e-a061-75e716522e77 req-7f2474da-f43d-4679-9e45-bfcd5d10434a service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Updating instance_info_cache with network_info: [{"id": "c063b2a3-1cf9-458a-a9ba-d3600501cd3e", "address": "fa:16:3e:75:c9:6b", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc063b2a3-1c", "ovs_interfaceid": "c063b2a3-1cf9-458a-a9ba-d3600501cd3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.165236] env[61906]: DEBUG oslo_concurrency.lockutils [req-62fd6042-4780-437e-a061-75e716522e77 req-7f2474da-f43d-4679-9e45-bfcd5d10434a service nova] Releasing lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.613281] env[61906]: DEBUG nova.compute.manager [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Start spawning the instance on the hypervisor. {{(pid=61906) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1136.642305] env[61906]: DEBUG nova.virt.hardware [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-24T14:02:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-24T14:02:39Z,direct_url=,disk_format='vmdk',id=29664dd5-796a-4dd3-b9f0-6a185cb0b7c1,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a4380a3e69a3476fa78c9cba75683c87',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-24T14:02:40Z,virtual_size=,visibility=), allow threads: False {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1136.642602] env[61906]: DEBUG nova.virt.hardware [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1136.642773] env[61906]: DEBUG nova.virt.hardware [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image limits 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1136.642967] env[61906]: DEBUG nova.virt.hardware [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Flavor pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1136.643138] env[61906]: DEBUG nova.virt.hardware [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Image pref 0:0:0 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1136.643295] env[61906]: DEBUG nova.virt.hardware [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61906) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1136.643509] env[61906]: DEBUG nova.virt.hardware [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1136.643719] env[61906]: DEBUG nova.virt.hardware [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1136.643913] env[61906]: DEBUG nova.virt.hardware [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Got 1 possible topologies {{(pid=61906) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1136.644098] env[61906]: DEBUG nova.virt.hardware [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1136.644281] env[61906]: DEBUG nova.virt.hardware [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61906) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1136.645162] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308219fd-f46c-4917-af84-63df5f9da962 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.653518] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5690e393-7f45-42d9-b553-07fd391d213f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.034976] env[61906]: DEBUG nova.compute.manager [req-26911b24-017e-4bbd-8315-5d4cfc094d88 req-02806e47-5d8e-4ec5-bbd3-e1c8526f37cb service nova] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Received event network-vif-plugged-6787aaee-7482-4d39-bed2-f88a99b55419 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1137.035430] env[61906]: DEBUG oslo_concurrency.lockutils [req-26911b24-017e-4bbd-8315-5d4cfc094d88 req-02806e47-5d8e-4ec5-bbd3-e1c8526f37cb service nova] Acquiring lock "660cbb4d-d924-4679-a017-a34c30b88c85-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.035843] env[61906]: DEBUG oslo_concurrency.lockutils [req-26911b24-017e-4bbd-8315-5d4cfc094d88 req-02806e47-5d8e-4ec5-bbd3-e1c8526f37cb service nova] Lock "660cbb4d-d924-4679-a017-a34c30b88c85-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.036225] env[61906]: DEBUG oslo_concurrency.lockutils [req-26911b24-017e-4bbd-8315-5d4cfc094d88 req-02806e47-5d8e-4ec5-bbd3-e1c8526f37cb service nova] Lock "660cbb4d-d924-4679-a017-a34c30b88c85-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.036581] env[61906]: DEBUG nova.compute.manager [req-26911b24-017e-4bbd-8315-5d4cfc094d88 req-02806e47-5d8e-4ec5-bbd3-e1c8526f37cb service nova] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] No waiting events found dispatching network-vif-plugged-6787aaee-7482-4d39-bed2-f88a99b55419 {{(pid=61906) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1137.036939] env[61906]: WARNING nova.compute.manager [req-26911b24-017e-4bbd-8315-5d4cfc094d88 req-02806e47-5d8e-4ec5-bbd3-e1c8526f37cb service nova] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Received unexpected event network-vif-plugged-6787aaee-7482-4d39-bed2-f88a99b55419 for instance with vm_state building and task_state spawning. [ 1137.120151] env[61906]: DEBUG nova.network.neutron [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Successfully updated port: 6787aaee-7482-4d39-bed2-f88a99b55419 {{(pid=61906) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1137.622576] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "refresh_cache-660cbb4d-d924-4679-a017-a34c30b88c85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.622903] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "refresh_cache-660cbb4d-d924-4679-a017-a34c30b88c85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.622903] env[61906]: DEBUG nova.network.neutron [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1138.157146] env[61906]: DEBUG nova.network.neutron [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Instance cache missing network info. {{(pid=61906) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1138.277088] env[61906]: DEBUG nova.network.neutron [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Updating instance_info_cache with network_info: [{"id": "6787aaee-7482-4d39-bed2-f88a99b55419", "address": "fa:16:3e:da:80:5e", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6787aaee-74", "ovs_interfaceid": "6787aaee-7482-4d39-bed2-f88a99b55419", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.779503] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "refresh_cache-660cbb4d-d924-4679-a017-a34c30b88c85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1138.779905] env[61906]: DEBUG nova.compute.manager [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Instance network_info: |[{"id": "6787aaee-7482-4d39-bed2-f88a99b55419", "address": "fa:16:3e:da:80:5e", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6787aaee-74", "ovs_interfaceid": "6787aaee-7482-4d39-bed2-f88a99b55419", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61906) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1138.780339] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:80:5e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1895250-76cc-41f7-b7f8-2e5679494607', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6787aaee-7482-4d39-bed2-f88a99b55419', 'vif_model': 'vmxnet3'}] {{(pid=61906) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1138.787765] env[61906]: DEBUG oslo.service.loopingcall [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1138.787976] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Creating VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1138.788222] env[61906]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac96eaa7-e771-42d1-880f-7c2e8cecec45 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.807671] env[61906]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1138.807671] env[61906]: value = "task-1357183" [ 1138.807671] env[61906]: _type = "Task" [ 1138.807671] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.815052] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357183, 'name': CreateVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.060493] env[61906]: DEBUG nova.compute.manager [req-39844cdb-3e91-4e00-aec6-7f06b1046741 req-35ef5e7a-11df-41b4-8029-3a0007738592 service nova] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Received event network-changed-6787aaee-7482-4d39-bed2-f88a99b55419 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1139.060824] env[61906]: DEBUG nova.compute.manager [req-39844cdb-3e91-4e00-aec6-7f06b1046741 req-35ef5e7a-11df-41b4-8029-3a0007738592 service nova] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Refreshing instance network info cache due to event network-changed-6787aaee-7482-4d39-bed2-f88a99b55419. {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1139.061045] env[61906]: DEBUG oslo_concurrency.lockutils [req-39844cdb-3e91-4e00-aec6-7f06b1046741 req-35ef5e7a-11df-41b4-8029-3a0007738592 service nova] Acquiring lock "refresh_cache-660cbb4d-d924-4679-a017-a34c30b88c85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.061243] env[61906]: DEBUG oslo_concurrency.lockutils [req-39844cdb-3e91-4e00-aec6-7f06b1046741 req-35ef5e7a-11df-41b4-8029-3a0007738592 service nova] Acquired lock "refresh_cache-660cbb4d-d924-4679-a017-a34c30b88c85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.061465] env[61906]: DEBUG nova.network.neutron [req-39844cdb-3e91-4e00-aec6-7f06b1046741 req-35ef5e7a-11df-41b4-8029-3a0007738592 service nova] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Refreshing network info cache for port 6787aaee-7482-4d39-bed2-f88a99b55419 {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1139.317609] env[61906]: DEBUG oslo_vmware.api [-] Task: {'id': task-1357183, 'name': CreateVM_Task, 'duration_secs': 0.281023} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.317755] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Created VM on the ESX host {{(pid=61906) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1139.318430] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.318599] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.318927] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1139.319194] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b8ea35b-c735-4d02-acaf-301098e53024 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.323769] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1139.323769] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c73991-9c0e-477f-a69f-c1ae7652c0a8" [ 1139.323769] env[61906]: _type = "Task" [ 1139.323769] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.332133] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c73991-9c0e-477f-a69f-c1ae7652c0a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.741965] env[61906]: DEBUG nova.network.neutron [req-39844cdb-3e91-4e00-aec6-7f06b1046741 req-35ef5e7a-11df-41b4-8029-3a0007738592 service nova] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Updated VIF entry in instance network info cache for port 6787aaee-7482-4d39-bed2-f88a99b55419. {{(pid=61906) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1139.742336] env[61906]: DEBUG nova.network.neutron [req-39844cdb-3e91-4e00-aec6-7f06b1046741 req-35ef5e7a-11df-41b4-8029-3a0007738592 service nova] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Updating instance_info_cache with network_info: [{"id": "6787aaee-7482-4d39-bed2-f88a99b55419", "address": "fa:16:3e:da:80:5e", "network": {"id": "7e2270d0-62a8-4cc6-aedc-5d4cf3203fa8", "bridge": "br-int", "label": "tempest-ServersTestJSON-722864282-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ac4f2c9065c46068c7837eb85cd7e20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6787aaee-74", "ovs_interfaceid": "6787aaee-7482-4d39-bed2-f88a99b55419", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.833837] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52c73991-9c0e-477f-a69f-c1ae7652c0a8, 'name': SearchDatastore_Task, 'duration_secs': 0.008223} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.834204] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.834383] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Processing image 29664dd5-796a-4dd3-b9f0-6a185cb0b7c1 {{(pid=61906) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1139.834619] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.834773] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquired lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.834956] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1139.835230] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2221157d-88de-4cd9-b8da-5c647cd2302f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.843215] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61906) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1139.843357] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61906) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1139.844052] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a003f468-6d4c-45d1-98be-92dc70fd0a61 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.848833] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1139.848833] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ef661f-6c2f-a4c9-b3ae-063441c657cc" [ 1139.848833] env[61906]: _type = "Task" [ 1139.848833] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.855826] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ef661f-6c2f-a4c9-b3ae-063441c657cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.245470] env[61906]: DEBUG oslo_concurrency.lockutils [req-39844cdb-3e91-4e00-aec6-7f06b1046741 req-35ef5e7a-11df-41b4-8029-3a0007738592 service nova] Releasing lock "refresh_cache-660cbb4d-d924-4679-a017-a34c30b88c85" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.358850] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52ef661f-6c2f-a4c9-b3ae-063441c657cc, 'name': SearchDatastore_Task, 'duration_secs': 0.007792} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.359640] env[61906]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca656cd6-1f69-4931-ba4f-b877a7b8485e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.364350] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1140.364350] env[61906]: value = "session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d53003-467a-5b4e-a014-a28e6e6ffbc0" [ 1140.364350] env[61906]: _type = "Task" [ 1140.364350] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.371196] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d53003-467a-5b4e-a014-a28e6e6ffbc0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.874643] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': session[52b49049-37b6-eb31-1da7-b5431cf8b0f2]52d53003-467a-5b4e-a014-a28e6e6ffbc0, 'name': SearchDatastore_Task, 'duration_secs': 0.008943} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.875036] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Releasing lock "[datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.875144] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 660cbb4d-d924-4679-a017-a34c30b88c85/660cbb4d-d924-4679-a017-a34c30b88c85.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1140.875399] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a01c164-e117-44f7-9f8b-6a0e61eff2bc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.881774] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1140.881774] env[61906]: value = "task-1357184" [ 1140.881774] env[61906]: _type = "Task" [ 1140.881774] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.888820] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357184, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.392451] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357184, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.893276] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357184, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514884} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.893599] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1/29664dd5-796a-4dd3-b9f0-6a185cb0b7c1.vmdk to [datastore1] 660cbb4d-d924-4679-a017-a34c30b88c85/660cbb4d-d924-4679-a017-a34c30b88c85.vmdk {{(pid=61906) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1141.893681] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Extending root virtual disk to 1048576 {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1141.893942] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55ae971b-b7bb-47a0-b459-a53f9a5d0505 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.900619] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1141.900619] env[61906]: value = "task-1357185" [ 1141.900619] env[61906]: _type = "Task" [ 1141.900619] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.908325] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357185, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.409910] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357185, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067027} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.410121] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Extended root virtual disk {{(pid=61906) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1142.410888] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0fb710-bc89-4a1b-ba2e-28fe061422dd {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.432137] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 660cbb4d-d924-4679-a017-a34c30b88c85/660cbb4d-d924-4679-a017-a34c30b88c85.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1142.432390] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7913fdf7-2a58-448c-9999-54fa1f6f0c4a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.450156] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1142.450156] env[61906]: value = "task-1357186" [ 1142.450156] env[61906]: _type = "Task" [ 1142.450156] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.458747] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357186, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.959762] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357186, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.459727] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357186, 'name': ReconfigVM_Task, 'duration_secs': 0.599926} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.460020] env[61906]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 660cbb4d-d924-4679-a017-a34c30b88c85/660cbb4d-d924-4679-a017-a34c30b88c85.vmdk or device None with type sparse {{(pid=61906) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1143.460632] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d920328d-df89-4024-9789-14346db8fef6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.466770] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1143.466770] env[61906]: value = "task-1357187" [ 1143.466770] env[61906]: _type = "Task" [ 1143.466770] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.478152] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357187, 'name': Rename_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.976125] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357187, 'name': Rename_Task, 'duration_secs': 0.125662} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.976485] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1143.976725] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f49bb1a4-0dd8-4e0c-b68c-60868206e294 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.982594] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1143.982594] env[61906]: value = "task-1357188" [ 1143.982594] env[61906]: _type = "Task" [ 1143.982594] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.989530] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357188, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.491934] env[61906]: DEBUG oslo_vmware.api [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357188, 'name': PowerOnVM_Task, 'duration_secs': 0.475893} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.492245] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1144.492456] env[61906]: INFO nova.compute.manager [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Took 7.88 seconds to spawn the instance on the hypervisor. [ 1144.492643] env[61906]: DEBUG nova.compute.manager [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1144.493433] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc955004-2f02-4697-af26-af0d39ba64b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.011725] env[61906]: INFO nova.compute.manager [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Took 12.58 seconds to build instance. [ 1145.514259] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4f30cc0d-c562-4fbd-9152-0cf5b144bc75 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "660cbb4d-d924-4679-a017-a34c30b88c85" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.085s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.769060] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7841752e-bddc-45ce-a9ae-bf394ae202bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "660cbb4d-d924-4679-a017-a34c30b88c85" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.769298] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7841752e-bddc-45ce-a9ae-bf394ae202bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "660cbb4d-d924-4679-a017-a34c30b88c85" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.769494] env[61906]: DEBUG nova.compute.manager [None req-7841752e-bddc-45ce-a9ae-bf394ae202bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1145.770459] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819731bc-4442-42f7-97b5-d625e34247aa {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.777356] env[61906]: DEBUG nova.compute.manager [None req-7841752e-bddc-45ce-a9ae-bf394ae202bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61906) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1145.777882] env[61906]: DEBUG nova.objects.instance [None req-7841752e-bddc-45ce-a9ae-bf394ae202bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lazy-loading 'flavor' on Instance uuid 660cbb4d-d924-4679-a017-a34c30b88c85 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1146.283164] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7841752e-bddc-45ce-a9ae-bf394ae202bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1146.283508] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2d12e83-63ae-4cdf-adf7-a13f5b2ef61c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.290935] env[61906]: DEBUG oslo_vmware.api [None req-7841752e-bddc-45ce-a9ae-bf394ae202bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1146.290935] env[61906]: value = "task-1357189" [ 1146.290935] env[61906]: _type = "Task" [ 1146.290935] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.298967] env[61906]: DEBUG oslo_vmware.api [None req-7841752e-bddc-45ce-a9ae-bf394ae202bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357189, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.800225] env[61906]: DEBUG oslo_vmware.api [None req-7841752e-bddc-45ce-a9ae-bf394ae202bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357189, 'name': PowerOffVM_Task, 'duration_secs': 0.192834} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.800452] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-7841752e-bddc-45ce-a9ae-bf394ae202bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1146.800633] env[61906]: DEBUG nova.compute.manager [None req-7841752e-bddc-45ce-a9ae-bf394ae202bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1146.801382] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5513c80b-16f5-4b0f-9d40-5980c38888e4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.312766] env[61906]: DEBUG oslo_concurrency.lockutils [None req-7841752e-bddc-45ce-a9ae-bf394ae202bc tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "660cbb4d-d924-4679-a017-a34c30b88c85" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.543s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.407055] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "660cbb4d-d924-4679-a017-a34c30b88c85" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.407460] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "660cbb4d-d924-4679-a017-a34c30b88c85" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.407552] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "660cbb4d-d924-4679-a017-a34c30b88c85-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.407776] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "660cbb4d-d924-4679-a017-a34c30b88c85-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.407963] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "660cbb4d-d924-4679-a017-a34c30b88c85-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.409987] env[61906]: INFO nova.compute.manager [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Terminating instance [ 1148.411639] env[61906]: DEBUG nova.compute.manager [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1148.411839] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1148.412661] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be50005-f4cd-4852-b63d-cd0d25b798a5 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.420252] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1148.420470] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3f8b7ab9-3645-4139-bf55-823b4f555fc7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.549332] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1148.549563] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1148.549782] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleting the datastore file [datastore1] 660cbb4d-d924-4679-a017-a34c30b88c85 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1148.550072] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97b8baff-79b2-43f3-ab52-9cb04504e938 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.555972] env[61906]: DEBUG oslo_vmware.api [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1148.555972] env[61906]: value = "task-1357191" [ 1148.555972] env[61906]: _type = "Task" [ 1148.555972] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.563259] env[61906]: DEBUG oslo_vmware.api [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357191, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.065840] env[61906]: DEBUG oslo_vmware.api [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357191, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.113339} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.066114] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1149.066307] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1149.066486] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1149.066938] env[61906]: INFO nova.compute.manager [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Took 0.65 seconds to destroy the instance on the hypervisor. [ 1149.066938] env[61906]: DEBUG oslo.service.loopingcall [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1149.067119] env[61906]: DEBUG nova.compute.manager [-] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1149.067215] env[61906]: DEBUG nova.network.neutron [-] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1149.300599] env[61906]: DEBUG nova.compute.manager [req-5bb7a342-618e-493c-939a-714931f8e6ea req-94dcde34-20f0-4b1d-9d4a-900ca35cdbb0 service nova] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Received event network-vif-deleted-6787aaee-7482-4d39-bed2-f88a99b55419 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1149.300782] env[61906]: INFO nova.compute.manager [req-5bb7a342-618e-493c-939a-714931f8e6ea req-94dcde34-20f0-4b1d-9d4a-900ca35cdbb0 service nova] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Neutron deleted interface 6787aaee-7482-4d39-bed2-f88a99b55419; detaching it from the instance and deleting it from the info cache [ 1149.300968] env[61906]: DEBUG nova.network.neutron [req-5bb7a342-618e-493c-939a-714931f8e6ea req-94dcde34-20f0-4b1d-9d4a-900ca35cdbb0 service nova] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.778295] env[61906]: DEBUG nova.network.neutron [-] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.803787] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b199d1fb-f666-451f-a217-8ce561a16305 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.815237] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f069ff-0e2e-49f8-ba15-bb8f431ceb93 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.837566] env[61906]: DEBUG nova.compute.manager [req-5bb7a342-618e-493c-939a-714931f8e6ea req-94dcde34-20f0-4b1d-9d4a-900ca35cdbb0 service nova] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Detach interface failed, port_id=6787aaee-7482-4d39-bed2-f88a99b55419, reason: Instance 660cbb4d-d924-4679-a017-a34c30b88c85 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1150.280879] env[61906]: INFO nova.compute.manager [-] [instance: 660cbb4d-d924-4679-a017-a34c30b88c85] Took 1.21 seconds to deallocate network for instance. [ 1150.787374] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.787770] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.787941] env[61906]: DEBUG nova.objects.instance [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lazy-loading 'resources' on Instance uuid 660cbb4d-d924-4679-a017-a34c30b88c85 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1151.341591] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103c0509-8f8c-4d47-ba71-e3739993b8fc {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.349281] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32801c6-65dc-4235-ad7b-8039f332a631 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.378133] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5b8984-f742-4eb4-81d8-c3ca4f4f1e3f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.384764] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c145bb-abcf-451b-bb46-9bcc33734d70 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.397142] env[61906]: DEBUG nova.compute.provider_tree [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1151.917171] env[61906]: ERROR nova.scheduler.client.report [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [req-e23b3a4f-d027-48d3-990d-f21278e2c068] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 6e524f5d-985e-41eb-a7ae-996be3ae1c1a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e23b3a4f-d027-48d3-990d-f21278e2c068"}]} [ 1151.933200] env[61906]: DEBUG nova.scheduler.client.report [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Refreshing inventories for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1151.945465] env[61906]: DEBUG nova.scheduler.client.report [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updating ProviderTree inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1151.945685] env[61906]: DEBUG nova.compute.provider_tree [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1151.956348] env[61906]: DEBUG nova.scheduler.client.report [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Refreshing aggregate associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, aggregates: None {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1151.973037] env[61906]: DEBUG nova.scheduler.client.report [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Refreshing trait associations for resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=61906) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1152.012034] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da59a45-1a3e-4e2e-962f-fe7ee22aacab {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.019036] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7e5eaf-5aa6-4d23-970a-a1b340dceccf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.048359] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a105b79f-ddd0-434d-96aa-cee33aba1114 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.054854] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67df5576-e8d7-4589-9810-cea7405bbeb9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.067232] env[61906]: DEBUG nova.compute.provider_tree [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1152.596437] env[61906]: DEBUG nova.scheduler.client.report [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updated inventory for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with generation 136 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1152.596704] env[61906]: DEBUG nova.compute.provider_tree [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updating resource provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a generation from 136 to 137 during operation: update_inventory {{(pid=61906) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1152.596889] env[61906]: DEBUG nova.compute.provider_tree [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Updating inventory in ProviderTree for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1153.102147] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.314s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.124628] env[61906]: INFO nova.scheduler.client.report [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted allocations for instance 660cbb4d-d924-4679-a017-a34c30b88c85 [ 1153.632903] env[61906]: DEBUG oslo_concurrency.lockutils [None req-0f6d93b5-0ec5-4d28-9c18-609e73e44761 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "660cbb4d-d924-4679-a017-a34c30b88c85" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.225s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.147796] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "91dc442c-0272-419d-8ef9-06e279a3ea61" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.148274] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "91dc442c-0272-419d-8ef9-06e279a3ea61" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.148625] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "91dc442c-0272-419d-8ef9-06e279a3ea61-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.148959] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "91dc442c-0272-419d-8ef9-06e279a3ea61-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.149282] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "91dc442c-0272-419d-8ef9-06e279a3ea61-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.152175] env[61906]: INFO nova.compute.manager [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Terminating instance [ 1154.154486] env[61906]: DEBUG nova.compute.manager [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1154.154737] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1154.155917] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ad646b-eba5-47c5-96b5-09ac5537389f {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.166233] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1154.166529] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f246538-e56c-4e6a-b7a3-b588b3ac485a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.173089] env[61906]: DEBUG oslo_vmware.api [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1154.173089] env[61906]: value = "task-1357192" [ 1154.173089] env[61906]: _type = "Task" [ 1154.173089] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.182969] env[61906]: DEBUG oslo_vmware.api [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357192, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.682648] env[61906]: DEBUG oslo_vmware.api [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357192, 'name': PowerOffVM_Task, 'duration_secs': 0.184494} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.682932] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1154.683127] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1154.683679] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7852667a-86f9-4c2a-956b-0698337b41a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.743404] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1154.743639] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Deleting contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1154.743834] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleting the datastore file [datastore2] 91dc442c-0272-419d-8ef9-06e279a3ea61 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1154.744114] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-29f2532d-0462-4535-b91b-f11e954d9433 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.750236] env[61906]: DEBUG oslo_vmware.api [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for the task: (returnval){ [ 1154.750236] env[61906]: value = "task-1357194" [ 1154.750236] env[61906]: _type = "Task" [ 1154.750236] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.757388] env[61906]: DEBUG oslo_vmware.api [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357194, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.260068] env[61906]: DEBUG oslo_vmware.api [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Task: {'id': task-1357194, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125545} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.260068] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1155.260497] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Deleted contents of the VM from datastore datastore2 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1155.260497] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1155.260497] env[61906]: INFO nova.compute.manager [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1155.260715] env[61906]: DEBUG oslo.service.loopingcall [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1155.260917] env[61906]: DEBUG nova.compute.manager [-] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1155.261017] env[61906]: DEBUG nova.network.neutron [-] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1155.479292] env[61906]: DEBUG nova.compute.manager [req-00c8d1f6-3261-4f2a-b815-3e6904308049 req-d56d2644-e08e-417f-a275-f81283efad49 service nova] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Received event network-vif-deleted-a10b72b0-6aab-4516-9e87-ce2ea1c1ae15 {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1155.479507] env[61906]: INFO nova.compute.manager [req-00c8d1f6-3261-4f2a-b815-3e6904308049 req-d56d2644-e08e-417f-a275-f81283efad49 service nova] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Neutron deleted interface a10b72b0-6aab-4516-9e87-ce2ea1c1ae15; detaching it from the instance and deleting it from the info cache [ 1155.479735] env[61906]: DEBUG nova.network.neutron [req-00c8d1f6-3261-4f2a-b815-3e6904308049 req-d56d2644-e08e-417f-a275-f81283efad49 service nova] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.958089] env[61906]: DEBUG nova.network.neutron [-] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.983364] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-54fc1f96-2eb1-4021-b071-a4d1173e3fd1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.992582] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ecce9c-0366-49d7-90e9-2cfa1194c255 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.013620] env[61906]: DEBUG nova.compute.manager [req-00c8d1f6-3261-4f2a-b815-3e6904308049 req-d56d2644-e08e-417f-a275-f81283efad49 service nova] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Detach interface failed, port_id=a10b72b0-6aab-4516-9e87-ce2ea1c1ae15, reason: Instance 91dc442c-0272-419d-8ef9-06e279a3ea61 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1156.464339] env[61906]: INFO nova.compute.manager [-] [instance: 91dc442c-0272-419d-8ef9-06e279a3ea61] Took 1.20 seconds to deallocate network for instance. [ 1156.970851] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.971182] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.971423] env[61906]: DEBUG nova.objects.instance [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lazy-loading 'resources' on Instance uuid 91dc442c-0272-419d-8ef9-06e279a3ea61 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1157.514802] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f70e6a94-afc6-4d39-ac57-943516ac9e0c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.522089] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfcdb745-9851-43ef-9ac5-390620e5902d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.551709] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba2e68e2-c22e-465f-8259-bb8d97d7d82c {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.558355] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c32d30c-c1da-43bd-92db-8841551b5865 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.570896] env[61906]: DEBUG nova.compute.provider_tree [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1158.075043] env[61906]: DEBUG nova.scheduler.client.report [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1158.578682] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.607s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.597185] env[61906]: INFO nova.scheduler.client.report [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Deleted allocations for instance 91dc442c-0272-419d-8ef9-06e279a3ea61 [ 1159.104489] env[61906]: DEBUG oslo_concurrency.lockutils [None req-d971b1e7-128f-4753-8934-4a960a8bb3e6 tempest-ServersTestJSON-1992257685 tempest-ServersTestJSON-1992257685-project-member] Lock "91dc442c-0272-419d-8ef9-06e279a3ea61" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.956s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.194385] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1169.194757] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1169.194804] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61906) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1170.194199] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1170.194440] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1171.194610] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1172.194302] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1172.194543] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1172.194712] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1172.325778] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f8e36089-68d5-4a1b-91ce-4e0d12348246 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.326078] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f8e36089-68d5-4a1b-91ce-4e0d12348246 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.326289] env[61906]: DEBUG nova.compute.manager [None req-f8e36089-68d5-4a1b-91ce-4e0d12348246 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1172.327205] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3f1ef2-1ed3-449e-8885-1a3f93e9b16d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.334626] env[61906]: DEBUG nova.compute.manager [None req-f8e36089-68d5-4a1b-91ce-4e0d12348246 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61906) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1172.335189] env[61906]: DEBUG nova.objects.instance [None req-f8e36089-68d5-4a1b-91ce-4e0d12348246 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lazy-loading 'flavor' on Instance uuid 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1172.697990] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.698262] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.698441] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.698600] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61906) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1172.699521] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84285eaa-ba2b-4f48-bbb9-68648a99cc91 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.707591] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64f957b3-9e3a-4415-9ab7-e2622044fbf1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.721155] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b18b6c1-ff67-433a-a104-82f37369b4a6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.727273] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b5ed09-7c3a-4791-9ff1-f588663a10f4 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.754681] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180925MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61906) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1172.754834] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.755025] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.840026] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8e36089-68d5-4a1b-91ce-4e0d12348246 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1172.840263] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d2025c73-3e74-4af3-bcc7-b527f6b25905 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.847514] env[61906]: DEBUG oslo_vmware.api [None req-f8e36089-68d5-4a1b-91ce-4e0d12348246 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1172.847514] env[61906]: value = "task-1357195" [ 1172.847514] env[61906]: _type = "Task" [ 1172.847514] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.856444] env[61906]: DEBUG oslo_vmware.api [None req-f8e36089-68d5-4a1b-91ce-4e0d12348246 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357195, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.356977] env[61906]: DEBUG oslo_vmware.api [None req-f8e36089-68d5-4a1b-91ce-4e0d12348246 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357195, 'name': PowerOffVM_Task, 'duration_secs': 0.173001} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.357369] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8e36089-68d5-4a1b-91ce-4e0d12348246 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1173.357490] env[61906]: DEBUG nova.compute.manager [None req-f8e36089-68d5-4a1b-91ce-4e0d12348246 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1173.358209] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9799fc-a0a5-4378-bf2b-124f5d24e015 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.777747] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Instance 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61906) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1173.777961] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1173.778122] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61906) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1173.802051] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651a92f8-0486-4a53-9387-7567608a781e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.809244] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e144b56-cab8-4217-a517-4610dd476094 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.838863] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f96aeb0-8ba2-46e5-98bb-b59dad5b7ac1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.845648] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2b83fd-2c02-43e4-a5b0-82a9e6fe1878 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.857905] env[61906]: DEBUG nova.compute.provider_tree [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1173.868776] env[61906]: DEBUG oslo_concurrency.lockutils [None req-f8e36089-68d5-4a1b-91ce-4e0d12348246 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.543s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.361626] env[61906]: DEBUG nova.scheduler.client.report [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1174.728405] env[61906]: DEBUG nova.objects.instance [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lazy-loading 'flavor' on Instance uuid 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1174.867711] env[61906]: DEBUG nova.compute.resource_tracker [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61906) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1174.867925] env[61906]: DEBUG oslo_concurrency.lockutils [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.113s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.234269] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.234501] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.234627] env[61906]: DEBUG nova.network.neutron [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1175.234811] env[61906]: DEBUG nova.objects.instance [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lazy-loading 'info_cache' on Instance uuid 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1175.738584] env[61906]: DEBUG nova.objects.base [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Object Instance<36f6adad-6df4-4e05-ba96-9fbfaaf38ad5> lazy-loaded attributes: flavor,info_cache {{(pid=61906) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1175.868446] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1175.868629] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Starting heal instance info cache {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1176.370831] env[61906]: DEBUG nova.compute.manager [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Didn't find any instances for network info cache update. {{(pid=61906) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1176.448858] env[61906]: DEBUG nova.network.neutron [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Updating instance_info_cache with network_info: [{"id": "c063b2a3-1cf9-458a-a9ba-d3600501cd3e", "address": "fa:16:3e:75:c9:6b", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc063b2a3-1c", "ovs_interfaceid": "c063b2a3-1cf9-458a-a9ba-d3600501cd3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.952162] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.455231] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Powering on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1177.455569] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1792bf0f-505d-418a-bf72-ecde8d77e06a {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.463699] env[61906]: DEBUG oslo_vmware.api [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1177.463699] env[61906]: value = "task-1357196" [ 1177.463699] env[61906]: _type = "Task" [ 1177.463699] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.471510] env[61906]: DEBUG oslo_vmware.api [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357196, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.691677] env[61906]: DEBUG oslo_service.periodic_task [None req-66f7ec57-222c-4dd8-a594-7437f5a873b9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61906) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1177.973946] env[61906]: DEBUG oslo_vmware.api [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357196, 'name': PowerOnVM_Task, 'duration_secs': 0.370617} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.974283] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Powered on the VM {{(pid=61906) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1177.974449] env[61906]: DEBUG nova.compute.manager [None req-4142ce17-7137-4034-8efa-84c2bc894b25 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1177.975211] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0bcbaa-54c1-4849-8984-acdf3c33ab16 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.910178] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792988bd-10e0-4168-a1b1-c9b82b0905a0 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.916245] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b8e9e2-9708-4620-acab-a0f5d9bbcf52 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Suspending the VM {{(pid=61906) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1179.916488] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-a02a36df-e01c-4791-9c12-aac1736e1fce {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.922985] env[61906]: DEBUG oslo_vmware.api [None req-f0b8e9e2-9708-4620-acab-a0f5d9bbcf52 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1179.922985] env[61906]: value = "task-1357197" [ 1179.922985] env[61906]: _type = "Task" [ 1179.922985] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.930198] env[61906]: DEBUG oslo_vmware.api [None req-f0b8e9e2-9708-4620-acab-a0f5d9bbcf52 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357197, 'name': SuspendVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.434825] env[61906]: DEBUG oslo_vmware.api [None req-f0b8e9e2-9708-4620-acab-a0f5d9bbcf52 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357197, 'name': SuspendVM_Task} progress is 75%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.933586] env[61906]: DEBUG oslo_vmware.api [None req-f0b8e9e2-9708-4620-acab-a0f5d9bbcf52 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357197, 'name': SuspendVM_Task, 'duration_secs': 0.55615} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.934059] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b8e9e2-9708-4620-acab-a0f5d9bbcf52 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Suspended the VM {{(pid=61906) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1180.934059] env[61906]: DEBUG nova.compute.manager [None req-f0b8e9e2-9708-4620-acab-a0f5d9bbcf52 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1180.934820] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e44646-8a3a-48c2-9462-715faff6852e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.227965] env[61906]: INFO nova.compute.manager [None req-74d6dc5b-3d9f-4de9-aac3-629a4bb99d66 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Resuming [ 1182.228611] env[61906]: DEBUG nova.objects.instance [None req-74d6dc5b-3d9f-4de9-aac3-629a4bb99d66 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lazy-loading 'flavor' on Instance uuid 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1183.236586] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74d6dc5b-3d9f-4de9-aac3-629a4bb99d66 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.237031] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74d6dc5b-3d9f-4de9-aac3-629a4bb99d66 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquired lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.237031] env[61906]: DEBUG nova.network.neutron [None req-74d6dc5b-3d9f-4de9-aac3-629a4bb99d66 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Building network info cache for instance {{(pid=61906) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1183.934509] env[61906]: DEBUG nova.network.neutron [None req-74d6dc5b-3d9f-4de9-aac3-629a4bb99d66 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Updating instance_info_cache with network_info: [{"id": "c063b2a3-1cf9-458a-a9ba-d3600501cd3e", "address": "fa:16:3e:75:c9:6b", "network": {"id": "07775921-5101-4d61-975d-26b729d78c1c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1221636620-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c6414fa082f49c2a568a772a256a8e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8233b50c-be21-441a-a357-a29138a95b8b", "external-id": "nsx-vlan-transportzone-496", "segmentation_id": 496, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc063b2a3-1c", "ovs_interfaceid": "c063b2a3-1cf9-458a-a9ba-d3600501cd3e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.437104] env[61906]: DEBUG oslo_concurrency.lockutils [None req-74d6dc5b-3d9f-4de9-aac3-629a4bb99d66 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Releasing lock "refresh_cache-36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" {{(pid=61906) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.438154] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e9bf50-2c37-49da-abe7-6d559bf54f06 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.445273] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-74d6dc5b-3d9f-4de9-aac3-629a4bb99d66 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Resuming the VM {{(pid=61906) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1184.445510] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9600e157-779c-47b0-84a6-a1e321d74baf {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.452225] env[61906]: DEBUG oslo_vmware.api [None req-74d6dc5b-3d9f-4de9-aac3-629a4bb99d66 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1184.452225] env[61906]: value = "task-1357198" [ 1184.452225] env[61906]: _type = "Task" [ 1184.452225] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.459994] env[61906]: DEBUG oslo_vmware.api [None req-74d6dc5b-3d9f-4de9-aac3-629a4bb99d66 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357198, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.963049] env[61906]: DEBUG oslo_vmware.api [None req-74d6dc5b-3d9f-4de9-aac3-629a4bb99d66 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357198, 'name': PowerOnVM_Task, 'duration_secs': 0.467612} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.963499] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-74d6dc5b-3d9f-4de9-aac3-629a4bb99d66 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Resumed the VM {{(pid=61906) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1184.963748] env[61906]: DEBUG nova.compute.manager [None req-74d6dc5b-3d9f-4de9-aac3-629a4bb99d66 tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Checking state {{(pid=61906) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1184.964674] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efde0835-f8d3-4752-9314-63ba2f5f190b {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.809349] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.809734] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.809895] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.810072] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.810262] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.812418] env[61906]: INFO nova.compute.manager [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Terminating instance [ 1185.814159] env[61906]: DEBUG nova.compute.manager [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Start destroying the instance on the hypervisor. {{(pid=61906) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1185.814358] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Destroying instance {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1185.815194] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821678d8-966d-4824-8022-55732bae43b1 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.823187] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Powering off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1185.823413] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c58dc48-46ea-4174-938c-01973a5ffe5e {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.829816] env[61906]: DEBUG oslo_vmware.api [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1185.829816] env[61906]: value = "task-1357199" [ 1185.829816] env[61906]: _type = "Task" [ 1185.829816] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.836867] env[61906]: DEBUG oslo_vmware.api [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357199, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.339993] env[61906]: DEBUG oslo_vmware.api [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357199, 'name': PowerOffVM_Task, 'duration_secs': 0.178284} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.340301] env[61906]: DEBUG nova.virt.vmwareapi.vm_util [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Powered off the VM {{(pid=61906) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1186.340475] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Unregistering the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1186.340724] env[61906]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-360930e2-6185-40dd-a346-887745a56459 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.404764] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Unregistered the VM {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1186.405273] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Deleting contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1186.405273] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleting the datastore file [datastore1] 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5 {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1186.405440] env[61906]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-401f0cef-3675-477b-9f73-6154953abfc9 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.411082] env[61906]: DEBUG oslo_vmware.api [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for the task: (returnval){ [ 1186.411082] env[61906]: value = "task-1357201" [ 1186.411082] env[61906]: _type = "Task" [ 1186.411082] env[61906]: } to complete. {{(pid=61906) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.418641] env[61906]: DEBUG oslo_vmware.api [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357201, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.920602] env[61906]: DEBUG oslo_vmware.api [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Task: {'id': task-1357201, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173887} completed successfully. {{(pid=61906) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.921025] env[61906]: DEBUG nova.virt.vmwareapi.ds_util [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleted the datastore file {{(pid=61906) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1186.921085] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Deleted contents of the VM from datastore datastore1 {{(pid=61906) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1186.921274] env[61906]: DEBUG nova.virt.vmwareapi.vmops [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Instance destroyed {{(pid=61906) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1186.921458] env[61906]: INFO nova.compute.manager [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1186.921699] env[61906]: DEBUG oslo.service.loopingcall [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61906) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1186.921895] env[61906]: DEBUG nova.compute.manager [-] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Deallocating network for instance {{(pid=61906) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1186.921991] env[61906]: DEBUG nova.network.neutron [-] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] deallocate_for_instance() {{(pid=61906) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1187.328454] env[61906]: DEBUG nova.compute.manager [req-89ae40ee-d864-4da7-85dd-6301696ddebe req-e75ba97b-b0d6-4a46-834c-9992ef86a228 service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Received event network-vif-deleted-c063b2a3-1cf9-458a-a9ba-d3600501cd3e {{(pid=61906) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1187.329350] env[61906]: INFO nova.compute.manager [req-89ae40ee-d864-4da7-85dd-6301696ddebe req-e75ba97b-b0d6-4a46-834c-9992ef86a228 service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Neutron deleted interface c063b2a3-1cf9-458a-a9ba-d3600501cd3e; detaching it from the instance and deleting it from the info cache [ 1187.329696] env[61906]: DEBUG nova.network.neutron [req-89ae40ee-d864-4da7-85dd-6301696ddebe req-e75ba97b-b0d6-4a46-834c-9992ef86a228 service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.806403] env[61906]: DEBUG nova.network.neutron [-] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Updating instance_info_cache with network_info: [] {{(pid=61906) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.832669] env[61906]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5af5572-02ed-4d64-a19b-186f03d557b6 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.842729] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94219135-3ef8-4390-810a-b6abd3d69525 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.865215] env[61906]: DEBUG nova.compute.manager [req-89ae40ee-d864-4da7-85dd-6301696ddebe req-e75ba97b-b0d6-4a46-834c-9992ef86a228 service nova] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Detach interface failed, port_id=c063b2a3-1cf9-458a-a9ba-d3600501cd3e, reason: Instance 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5 could not be found. {{(pid=61906) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1188.309288] env[61906]: INFO nova.compute.manager [-] [instance: 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5] Took 1.39 seconds to deallocate network for instance. [ 1188.819109] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.819109] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.819109] env[61906]: DEBUG nova.objects.instance [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lazy-loading 'resources' on Instance uuid 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5 {{(pid=61906) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.351731] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5fd9025-83ee-4ee3-a36d-cf210b1c25d7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.359201] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b7c38b-2aff-4e41-9ccd-b2fd27338044 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.387926] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b9087e-a446-4d9e-9e6c-1f07d62a095d {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.395648] env[61906]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e12378-fa66-4ad1-951e-c1e3e9c480e7 {{(pid=61906) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.408426] env[61906]: DEBUG nova.compute.provider_tree [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed in ProviderTree for provider: 6e524f5d-985e-41eb-a7ae-996be3ae1c1a {{(pid=61906) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1189.911906] env[61906]: DEBUG nova.scheduler.client.report [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Inventory has not changed for provider 6e524f5d-985e-41eb-a7ae-996be3ae1c1a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61906) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1190.418298] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.600s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.439872] env[61906]: INFO nova.scheduler.client.report [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Deleted allocations for instance 36f6adad-6df4-4e05-ba96-9fbfaaf38ad5 [ 1190.949598] env[61906]: DEBUG oslo_concurrency.lockutils [None req-4af4aed5-07c7-48ad-a833-d485300ca5bf tempest-ServerActionsTestJSON-2105777202 tempest-ServerActionsTestJSON-2105777202-project-member] Lock "36f6adad-6df4-4e05-ba96-9fbfaaf38ad5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.140s {{(pid=61906) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}